首页 > 最新文献

2009 Digital Image Computing: Techniques and Applications最新文献

英文 中文
Automatic Detection of Defective Zebrafish Embryos via Shape Analysis 斑马鱼胚胎缺陷的形状分析自动检测
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.76
Haifeng Zhao, J. Zhou, A. Robles-Kelly, Jianfeng Lu, Jing-yu Yang
In this paper, we present a graph-based approach to automatically detect defective zebrafish embryos. Here, the zebrafish is segmented from the background using a texture descriptor and morphological operations. In this way, we can represent the embryo shape as a graph, for which we propose a vectorisation method to recover clique histogram vectors for classification. The clique histogram represents the distribution of one vertex with respect to its adjacent vertices. This treatment permits the use of a codebook approach to represent the graph in terms of a set of codewords that can be used for purposes of support vector machine classification. The experimental results show that the method is not only effective but also robust to occlusions and shape variations. represent the embryo shape as a graph, for which we propose a vectorisation method to recover clique histogram vectors for classification. The clique histogram represents the distribution of one vertex with respect to its adjacent vertices. This treatment permits the use of a codebook approach to represent the graph in terms of a set of codewords that can be used for purposes of support vector machine classification. The experimental results show that the method is not only effective but also robust to occlusions and shape variations.
在本文中,我们提出了一种基于图的方法来自动检测缺陷斑马鱼胚胎。在这里,使用纹理描述符和形态学操作将斑马鱼从背景中分割出来。这样,我们可以将胚胎形状表示为一个图,为此我们提出了一种矢量化方法来恢复团直方图向量进行分类。团直方图表示一个顶点相对于其相邻顶点的分布。这种处理允许使用码本方法来表示一组码字,这些码字可用于支持向量机分类。实验结果表明,该方法不仅有效,而且对遮挡和形状变化具有较强的鲁棒性。将胚胎形状表示为图形,提出了一种矢量化方法来恢复团直方图向量进行分类。团直方图表示一个顶点相对于其相邻顶点的分布。这种处理允许使用码本方法来表示一组码字,这些码字可用于支持向量机分类。实验结果表明,该方法不仅有效,而且对遮挡和形状变化具有较强的鲁棒性。
{"title":"Automatic Detection of Defective Zebrafish Embryos via Shape Analysis","authors":"Haifeng Zhao, J. Zhou, A. Robles-Kelly, Jianfeng Lu, Jing-yu Yang","doi":"10.1109/DICTA.2009.76","DOIUrl":"https://doi.org/10.1109/DICTA.2009.76","url":null,"abstract":"In this paper, we present a graph-based approach to automatically detect defective zebrafish embryos. Here, the zebrafish is segmented from the background using a texture descriptor and morphological operations. In this way, we can represent the embryo shape as a graph, for which we propose a vectorisation method to recover clique histogram vectors for classification. The clique histogram represents the distribution of one vertex with respect to its adjacent vertices. This treatment permits the use of a codebook approach to represent the graph in terms of a set of codewords that can be used for purposes of support vector machine classification. The experimental results show that the method is not only effective but also robust to occlusions and shape variations. represent the embryo shape as a graph, for which we propose a vectorisation method to recover clique histogram vectors for classification. The clique histogram represents the distribution of one vertex with respect to its adjacent vertices. This treatment permits the use of a codebook approach to represent the graph in terms of a set of codewords that can be used for purposes of support vector machine classification. The experimental results show that the method is not only effective but also robust to occlusions and shape variations.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129957491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Measuring Latency for Video Surveillance Systems 视频监控系统的延迟测量
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.23
R. Hill, Christopher S. Madden, A. Hengel, Henry Detmold, A. Dick
The increased flexibility and other benefits offered by IP network cameras makes them a common choice for installation in new and expanded surveillance networks. One commonly quoted limitation of IP cameras is their high latency when compared to their analogue counterparts. This causes some reluctance to install or upgrade to digital cameras, and is slowing the adoption of live, intelligent analysis techniques in video surveillance systems. This paper presents methods for measurement of the latency in systems based upon digital IP or analogue cameras. These methods are camera-agnostic and require no specialised hardware. We use these methods to compare a variety of camera models. The results demonstrate that whilst analogue cameras do have a lower latency, most IP cameras are within acceptable tolerances. The source of the latency within an IP camera is also analysed, with prospects for improvement identified.
IP网络摄像机增加的灵活性和其他好处使它们成为安装在新的和扩展的监控网络中的常见选择。与模拟摄像机相比,IP摄像机的一个常见限制是其高延迟。这导致一些人不愿意安装或升级到数码相机,并减缓了视频监控系统中实时智能分析技术的采用。本文介绍了基于数字IP或模拟摄像机的系统延迟测量方法。这些方法与相机无关,不需要专门的硬件。我们使用这些方法来比较各种相机模型。结果表明,虽然模拟摄像机确实具有较低的延迟,但大多数IP摄像机都在可接受的公差范围内。还分析了IP摄像机内延迟的来源,并确定了改进的前景。
{"title":"Measuring Latency for Video Surveillance Systems","authors":"R. Hill, Christopher S. Madden, A. Hengel, Henry Detmold, A. Dick","doi":"10.1109/DICTA.2009.23","DOIUrl":"https://doi.org/10.1109/DICTA.2009.23","url":null,"abstract":"The increased flexibility and other benefits offered by IP network cameras makes them a common choice for installation in new and expanded surveillance networks. One commonly quoted limitation of IP cameras is their high latency when compared to their analogue counterparts. This causes some reluctance to install or upgrade to digital cameras, and is slowing the adoption of live, intelligent analysis techniques in video surveillance systems. This paper presents methods for measurement of the latency in systems based upon digital IP or analogue cameras. These methods are camera-agnostic and require no specialised hardware. We use these methods to compare a variety of camera models. The results demonstrate that whilst analogue cameras do have a lower latency, most IP cameras are within acceptable tolerances. The source of the latency within an IP camera is also analysed, with prospects for improvement identified.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125474805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
3D Reconstruction of Patient Specific Bone Models from 2D Radiographs for Image Guided Orthopedic Surgery 图像引导骨科手术中2D x线片患者特异性骨模型的3D重建
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.42
P. Gamage, S. Xie, P. Delmas, P. Xu
Three dimensional (3D) visualization of anatomy plays an important role in image guided orthopedic surgery and ultimately motivates minimally invasive procedures. However, direct 3D imaging modalities such as Computed Tomography (CT) are restricted to a minority of complex orthopedic procedures. Thus the diagnostics and planning of many interventions still rely on two dimensional (2D) radiographic images, where the surgeon has to mentally visualize the anatomy of interest. The purpose of this paper is to apply and validate a bi-planar 3D reconstruction methodology driven by prominent bony anatomy edges and contours identified on orthogonal radiographs. The results obtained through the proposed methodology are benchmarked against 3D CT scan data to assess the accuracy of reconstruction. The human femur has been used as the anatomy of interest throughout the paper. The novelty of this methodology is that it not only involves the outer contours of the bony anatomy in the reconstruction but also several key interior edges identifiable on radiographic images. Hence, this framework is not simply limited to long bones, but is generally applicable to a multitude of other bony anatomies as illustrated in the results section.
三维解剖可视化在图像引导骨科手术中起着重要作用,并最终推动微创手术的发展。然而,直接的三维成像模式,如计算机断层扫描(CT)仅限于少数复杂的骨科手术。因此,许多干预的诊断和计划仍然依赖于二维(2D)放射图像,外科医生必须在心理上可视化感兴趣的解剖结构。本文的目的是应用和验证一种双平面三维重建方法,该方法由正交x线片上识别的突出骨骼解剖边缘和轮廓驱动。通过所提出的方法获得的结果与3D CT扫描数据进行基准测试,以评估重建的准确性。人类股骨一直被用作整个论文感兴趣的解剖结构。这种方法的新颖之处在于,它不仅涉及重建骨骼解剖的外部轮廓,而且还涉及在放射图像上可识别的几个关键的内部边缘。因此,这个框架不仅限于长骨,而且通常适用于结果部分所示的许多其他骨骼解剖结构。
{"title":"3D Reconstruction of Patient Specific Bone Models from 2D Radiographs for Image Guided Orthopedic Surgery","authors":"P. Gamage, S. Xie, P. Delmas, P. Xu","doi":"10.1109/DICTA.2009.42","DOIUrl":"https://doi.org/10.1109/DICTA.2009.42","url":null,"abstract":"Three dimensional (3D) visualization of anatomy plays an important role in image guided orthopedic surgery and ultimately motivates minimally invasive procedures. However, direct 3D imaging modalities such as Computed Tomography (CT) are restricted to a minority of complex orthopedic procedures. Thus the diagnostics and planning of many interventions still rely on two dimensional (2D) radiographic images, where the surgeon has to mentally visualize the anatomy of interest. The purpose of this paper is to apply and validate a bi-planar 3D reconstruction methodology driven by prominent bony anatomy edges and contours identified on orthogonal radiographs. The results obtained through the proposed methodology are benchmarked against 3D CT scan data to assess the accuracy of reconstruction. The human femur has been used as the anatomy of interest throughout the paper. The novelty of this methodology is that it not only involves the outer contours of the bony anatomy in the reconstruction but also several key interior edges identifiable on radiographic images. Hence, this framework is not simply limited to long bones, but is generally applicable to a multitude of other bony anatomies as illustrated in the results section.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114506456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Dense Correspondence Extraction in Difficult Uncalibrated Scenarios 在困难的非校准场景密集对应提取
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.19
R. Lakemond, C. Fookes, S. Sridharan
The relationship between multiple cameras viewing the same scene may be discovered automatically by finding corresponding points in the two views and then solving for the camera geometry. In camera networks with sparsely placed cameras, low resolution cameras or in scenes with few distinguishable features it may be difficult to find a sufficient number of reliable correspondences from which to compute geometry. This paper presents a method for extracting a larger number of correspondences from an initial set of putative correspondences without any knowledge of the scene or camera geometry. The method may be used to increase the number of correspondences and make geometry computations possible in cases where existing methods have produced insufficient correspondences.
通过在两个视图中找到对应的点,然后求解相机几何,可以自动发现观看同一场景的多个相机之间的关系。在摄像机网络中,摄像机的位置稀疏,低分辨率摄像机或场景中几乎没有可区分的特征,可能很难找到足够数量的可靠对应来计算几何。本文提出了一种从一组初始假定对应中提取大量对应的方法,而无需了解场景或相机几何形状。该方法可用于增加对应的数量,并在现有方法产生的对应不足的情况下使几何计算成为可能。
{"title":"Dense Correspondence Extraction in Difficult Uncalibrated Scenarios","authors":"R. Lakemond, C. Fookes, S. Sridharan","doi":"10.1109/DICTA.2009.19","DOIUrl":"https://doi.org/10.1109/DICTA.2009.19","url":null,"abstract":"The relationship between multiple cameras viewing the same scene may be discovered automatically by finding corresponding points in the two views and then solving for the camera geometry. In camera networks with sparsely placed cameras, low resolution cameras or in scenes with few distinguishable features it may be difficult to find a sufficient number of reliable correspondences from which to compute geometry. This paper presents a method for extracting a larger number of correspondences from an initial set of putative correspondences without any knowledge of the scene or camera geometry. The method may be used to increase the number of correspondences and make geometry computations possible in cases where existing methods have produced insufficient correspondences.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117068660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Handling Significant Scale Difference for Object Retrieval in a Supermarket 超市对象检索的显著尺度差异处理
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.79
Yuhang Zhang, Lei Wang, R. Hartley, Hongdong Li
We propose an object retrieval application which can retrieve user specified objects from a big supermarket. Significant and unpredictable scale difference between the query and the database image is the major obstacle encountered. The widely used local invariant features show their deficiency in such an occasion. To improve the situation, we first design a new weighting scheme which can assess the repeatability of local features against scale variance. Also, another method which deals with scale difference through retrieving a query under multiple scales is also developed. Our methods have been tested on a real image database collected from a local supermarket and outperform the existing local invariant feature based image retrieval approaches. A new spatial check method is also briefly discussed.
提出了一个对象检索应用程序,可以从大型超市中检索用户指定的对象。查询和数据库映像之间显著且不可预测的规模差异是遇到的主要障碍。在这种情况下,广泛使用的局部不变量特征显示出它们的不足。为了改善这种情况,我们首先设计了一种新的加权方案,该方案可以根据尺度方差评估局部特征的可重复性。此外,本文还提出了一种通过在多个尺度下检索查询来处理尺度差异的方法。我们的方法已经在从本地超市收集的真实图像数据库上进行了测试,并且优于现有的基于局部不变特征的图像检索方法。并简要讨论了一种新的空间校核方法。
{"title":"Handling Significant Scale Difference for Object Retrieval in a Supermarket","authors":"Yuhang Zhang, Lei Wang, R. Hartley, Hongdong Li","doi":"10.1109/DICTA.2009.79","DOIUrl":"https://doi.org/10.1109/DICTA.2009.79","url":null,"abstract":"We propose an object retrieval application which can retrieve user specified objects from a big supermarket. Significant and unpredictable scale difference between the query and the database image is the major obstacle encountered. The widely used local invariant features show their deficiency in such an occasion. To improve the situation, we first design a new weighting scheme which can assess the repeatability of local features against scale variance. Also, another method which deals with scale difference through retrieving a query under multiple scales is also developed. Our methods have been tested on a real image database collected from a local supermarket and outperform the existing local invariant feature based image retrieval approaches. A new spatial check method is also briefly discussed.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"611 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123322043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Modeling Image Context Using Object Centered Grid 使用对象中心网格建模图像上下文
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.80
S. N. Parizi, I. Laptev, Alireza Tavakoli Targhi
Context plays a valuable role in any image understanding task confirmed by numerous studies which have shown the importance of contextual information in computer vision tasks, like object detection, scene classification and image retrieval. Studies of human perception on the tasks of scene classification and visual search have shown that human visual system makes extensive use of contextual information as postprocessing in order to index objects. Several recent computer vision approaches use contextual information to improve object recognition performance. They mainly use global information of the whole image by dividing the image into several predefined subregions, so called fixed grid. In this paper we propose an alternative approach to retrieval of contextual information, by customizing the location of the grid based on salient objects in the image. We claim this approach to result in more informative contextual features compared to the fixed grid based strategy. To compare our results with the most relevant and recent papers, we use PASCAL 2007 data set. Our experimental results show an improvement in terms of Mean Average Precision.
上下文在任何图像理解任务中都起着重要的作用,许多研究都证实了上下文信息在计算机视觉任务中的重要性,如物体检测、场景分类和图像检索。人类对场景分类和视觉搜索任务的感知研究表明,人类视觉系统广泛使用上下文信息作为后处理来索引对象。最近的几种计算机视觉方法使用上下文信息来提高对象识别性能。它们主要利用整个图像的全局信息,将图像划分为几个预定义的子区域,即固定网格。在本文中,我们提出了一种替代方法来检索上下文信息,通过自定义网格的位置基于图像中的显著对象。我们声称,与基于固定网格的策略相比,这种方法可以产生更多信息丰富的上下文特征。为了将我们的结果与最相关和最新的论文进行比较,我们使用了PASCAL 2007数据集。实验结果表明,该方法在平均精度方面有所提高。
{"title":"Modeling Image Context Using Object Centered Grid","authors":"S. N. Parizi, I. Laptev, Alireza Tavakoli Targhi","doi":"10.1109/DICTA.2009.80","DOIUrl":"https://doi.org/10.1109/DICTA.2009.80","url":null,"abstract":"Context plays a valuable role in any image understanding task confirmed by numerous studies which have shown the importance of contextual information in computer vision tasks, like object detection, scene classification and image retrieval. Studies of human perception on the tasks of scene classification and visual search have shown that human visual system makes extensive use of contextual information as postprocessing in order to index objects. Several recent computer vision approaches use contextual information to improve object recognition performance. They mainly use global information of the whole image by dividing the image into several predefined subregions, so called fixed grid. In this paper we propose an alternative approach to retrieval of contextual information, by customizing the location of the grid based on salient objects in the image. We claim this approach to result in more informative contextual features compared to the fixed grid based strategy. To compare our results with the most relevant and recent papers, we use PASCAL 2007 data set. Our experimental results show an improvement in terms of Mean Average Precision.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122995682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Learning the Optimal Transformation of Salient Features for Image Classification 学习图像分类中显著特征的最优变换
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.28
J. Zhou, Zhouyu Fu, A. Robles-Kelly
In this paper, we address the problem of recovering an optimal salient image descriptor transformation for image classification. Our method involves two steps. Firstly, a binary salient map is generated to specify the regions of interest for subsequent image feature extraction. To this end, an optimal cut-off value is recovered by maximising Fisher’s linear discriminant separability measure so as to separate the salient regions from the background of the scene. Next, image descriptors are extracted in the foreground region in order to be optimally transformed. The descriptor optimisation problem is cast in a regularised risk minimisation setting, in which the aim of computation is to recover the optimal transformation up to a cost function. The cost function is convex and can be solved using quadratic programming. The results on unsegmented Oxford Flowers database show that the proposed method can achieve classification performance that are comparable to those provided by alternatives elsewhere in the literature which employ pre-segmented images.
在本文中,我们解决了恢复图像分类的最优显著图像描述子变换的问题。我们的方法包括两个步骤。首先,生成二值显著图,指定感兴趣的区域,用于随后的图像特征提取。为此,通过最大化Fisher的线性判别可分离性度量来恢复最佳截止值,从而将突出区域从场景的背景中分离出来。接下来,在前景区域提取图像描述符,以便进行最优变换。描述符优化问题是在一个正则化的风险最小化设置中进行的,其中计算的目的是恢复到成本函数的最优转换。代价函数是凸的,可以用二次规划求解。在未分割的Oxford Flowers数据库上的结果表明,所提出的方法可以达到与文献中使用预分割图像的替代方法相当的分类性能。
{"title":"Learning the Optimal Transformation of Salient Features for Image Classification","authors":"J. Zhou, Zhouyu Fu, A. Robles-Kelly","doi":"10.1109/DICTA.2009.28","DOIUrl":"https://doi.org/10.1109/DICTA.2009.28","url":null,"abstract":"In this paper, we address the problem of recovering an optimal salient image descriptor transformation for image classification. Our method involves two steps. Firstly, a binary salient map is generated to specify the regions of interest for subsequent image feature extraction. To this end, an optimal cut-off value is recovered by maximising Fisher’s linear discriminant separability measure so as to separate the salient regions from the background of the scene. Next, image descriptors are extracted in the foreground region in order to be optimally transformed. The descriptor optimisation problem is cast in a regularised risk minimisation setting, in which the aim of computation is to recover the optimal transformation up to a cost function. The cost function is convex and can be solved using quadratic programming. The results on unsegmented Oxford Flowers database show that the proposed method can achieve classification performance that are comparable to those provided by alternatives elsewhere in the literature which employ pre-segmented images.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122937586","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Investigations into the Robustness of Audio-Visual Gender Classification to Background Noise and Illumination Effects 视听性别分类对背景噪声和光照影响的鲁棒性研究
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.34
D. Stewart, Hongbin Wang, Jiali Shen, P. Miller
In this paper we investigate the robustness of a multimodal gender profiling system which uses face and voice modalities. We use support vector machines combined with principal component analysis features to model faces, and Gaussian mixture models with Mel Frequency Cepstral Coefficients to model voices. Our results show that these approaches perform well individually in ‘clean’ training and testing conditions but that their performance can deteriorate substantially in the presence of audio or image corruptions such as additive acoustic noise and differing image illumination conditions. However, our results also show that a straightforward combination of these modalities can provide a gender classifier which is robust when tested in the presence of corruption in either modality. We also show that in most of the tested conditions the multimodal system can automatically perform on a par with whichever single modality is currently the most reliable.
在本文中,我们研究了多模态性别分析系统的鲁棒性,该系统使用面部和声音模式。我们使用结合主成分分析特征的支持向量机来建模人脸,使用带有Mel频率倒谱系数的高斯混合模型来建模声音。我们的研究结果表明,这些方法在“干净”的训练和测试条件下单独表现良好,但在存在音频或图像损坏(如附加噪声和不同的图像照明条件)的情况下,它们的性能可能会大幅下降。然而,我们的结果也表明,这些模式的直接组合可以提供一个性别分类器,当在任何一种模式存在腐败的情况下进行测试时,该分类器都是稳健的。我们还表明,在大多数测试条件下,多模态系统可以自动执行与当前最可靠的单模态相同的性能。
{"title":"Investigations into the Robustness of Audio-Visual Gender Classification to Background Noise and Illumination Effects","authors":"D. Stewart, Hongbin Wang, Jiali Shen, P. Miller","doi":"10.1109/DICTA.2009.34","DOIUrl":"https://doi.org/10.1109/DICTA.2009.34","url":null,"abstract":"In this paper we investigate the robustness of a multimodal gender profiling system which uses face and voice modalities. We use support vector machines combined with principal component analysis features to model faces, and Gaussian mixture models with Mel Frequency Cepstral Coefficients to model voices. Our results show that these approaches perform well individually in ‘clean’ training and testing conditions but that their performance can deteriorate substantially in the presence of audio or image corruptions such as additive acoustic noise and differing image illumination conditions. However, our results also show that a straightforward combination of these modalities can provide a gender classifier which is robust when tested in the presence of corruption in either modality. We also show that in most of the tested conditions the multimodal system can automatically perform on a par with whichever single modality is currently the most reliable.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123969753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Improved Simultaneous Computation of Motion Detection and Optical Flow for Object Tracking 改进的运动检测和光流同步计算的目标跟踪
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.35
S. Denman, C. Fookes, S. Sridharan
Object tracking systems require accurate segmentation of the objects from the background for effective tracking. Motion segmentation or optical flow can be used to segment incoming images. Whilst optical flow allows multiple moving targets to be separated based on their individual velocities, optical flow techniques are prone to errors caused by changing lighting and occlusions, both common in a surveillance environment. Motion segmentation techniques are more robust to fluctuating lighting and occlusions, but don't provide information on the direction of the motion. In this paper we propose a combined motion segmentation/optical flow algorithm for use in object tracking. The proposed algorithm uses the motion segmentation results to inform the optical flow calculations and ensure that optical flow is only calculated in regions of motion, and improve the performance of the optical flow around the edge of moving objects. Optical flow is calculated at pixel resolution and tracking of flow vectors is employed to improve performance and detect discontinuities, which can indicate the location of overlaps between objects. The algorithm is evaluated by attempting to extract a moving target within the flow images, given expected horizontal and vertical movement (i.e. the algorithms intended use for object tracking). Results show that the proposed algorithm outperforms other widely used optical flow techniques for this surveillance application.
目标跟踪系统需要准确地从背景中分割目标以实现有效的跟踪。运动分割或光流可用于分割传入图像。虽然光流允许根据单个速度分离多个移动目标,但光流技术容易受到光照和遮挡变化引起的误差,这在监视环境中都很常见。运动分割技术对波动光照和遮挡更加稳健,但不提供运动方向的信息。本文提出了一种用于目标跟踪的运动分割/光流组合算法。该算法利用运动分割结果进行光流计算,保证光流只在运动区域进行计算,提高了运动物体边缘周围光流的性能。以像素分辨率计算光流,并采用流矢量跟踪来提高性能和检测不连续点,这可以指示物体之间重叠的位置。该算法通过尝试提取流图像中的运动目标来评估,给定预期的水平和垂直运动(即用于对象跟踪的算法)。结果表明,该算法优于其他广泛使用的光流监控技术。
{"title":"Improved Simultaneous Computation of Motion Detection and Optical Flow for Object Tracking","authors":"S. Denman, C. Fookes, S. Sridharan","doi":"10.1109/DICTA.2009.35","DOIUrl":"https://doi.org/10.1109/DICTA.2009.35","url":null,"abstract":"Object tracking systems require accurate segmentation of the objects from the background for effective tracking. Motion segmentation or optical flow can be used to segment incoming images. Whilst optical flow allows multiple moving targets to be separated based on their individual velocities, optical flow techniques are prone to errors caused by changing lighting and occlusions, both common in a surveillance environment. Motion segmentation techniques are more robust to fluctuating lighting and occlusions, but don't provide information on the direction of the motion. In this paper we propose a combined motion segmentation/optical flow algorithm for use in object tracking. The proposed algorithm uses the motion segmentation results to inform the optical flow calculations and ensure that optical flow is only calculated in regions of motion, and improve the performance of the optical flow around the edge of moving objects. Optical flow is calculated at pixel resolution and tracking of flow vectors is employed to improve performance and detect discontinuities, which can indicate the location of overlaps between objects. The algorithm is evaluated by attempting to extract a moving target within the flow images, given expected horizontal and vertical movement (i.e. the algorithms intended use for object tracking). Results show that the proposed algorithm outperforms other widely used optical flow techniques for this surveillance application.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128392319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 69
Straight-Edge Extraction in Distorted Images Using Gradient Correction 基于梯度校正的畸变图像的直边提取
Pub Date : 2009-12-01 DOI: 10.1109/DICTA.2009.86
M. Islam, L. Kitchen
Many camera lenses, particularly low-cost or wide-angle lenses, can cause significant image distortion. This means that features extracted naively from such images will be incorrect. A traditional approach to dealing with this problem is to digitally rectify the image to correct the distortion, and then to apply computer vision processing to the corrected image. However, this is relatively expensive computationally, and can introduce additional interpolation errors. We propose instead to apply processing directly to the distorted image from the camera, modifying whatever algorithm is used to correct for the distortion during processing, without a separate rectification pass. In this paper we demonstrate the effectiveness of this approach using the particular classic problem of gradient-based extraction of straight edges. We propose a modification of the Burns line extractor that works on a distorted image by correcting the gradients on the fly using the chain rule, and correcting the pixel positions during the line-fitting stage. Experimental results on both real and synthetic images under varying distortion and noise show that our gradient-correction technique can obtain approximately a 50% reduction in computation time for straight-edge extraction, with a modest improvement in accuracy under most conditions.
许多相机镜头,特别是低成本或广角镜头,会造成严重的图像失真。这意味着从这些图像中天真地提取的特征将是不正确的。处理这一问题的传统方法是对图像进行数字校正以校正畸变,然后对校正后的图像进行计算机视觉处理。然而,这是相对昂贵的计算,并可能引入额外的插值误差。我们建议直接对来自相机的扭曲图像进行处理,在处理过程中修改用于纠正扭曲的任何算法,而不需要单独的校正通道。在本文中,我们用基于梯度的直线边提取的经典问题证明了这种方法的有效性。我们提出了对Burns线提取器的改进,该方法通过使用链式法则在动态中校正梯度,并在线拟合阶段校正像素位置,从而对扭曲图像起作用。在不同失真和噪声的真实图像和合成图像上的实验结果表明,我们的梯度校正技术可以使直线边缘提取的计算时间减少大约50%,在大多数情况下精度都有适度的提高。
{"title":"Straight-Edge Extraction in Distorted Images Using Gradient Correction","authors":"M. Islam, L. Kitchen","doi":"10.1109/DICTA.2009.86","DOIUrl":"https://doi.org/10.1109/DICTA.2009.86","url":null,"abstract":"Many camera lenses, particularly low-cost or wide-angle lenses, can cause significant image distortion. This means that features extracted naively from such images will be incorrect. A traditional approach to dealing with this problem is to digitally rectify the image to correct the distortion, and then to apply computer vision processing to the corrected image. However, this is relatively expensive computationally, and can introduce additional interpolation errors. We propose instead to apply processing directly to the distorted image from the camera, modifying whatever algorithm is used to correct for the distortion during processing, without a separate rectification pass. In this paper we demonstrate the effectiveness of this approach using the particular classic problem of gradient-based extraction of straight edges. We propose a modification of the Burns line extractor that works on a distorted image by correcting the gradients on the fly using the chain rule, and correcting the pixel positions during the line-fitting stage. Experimental results on both real and synthetic images under varying distortion and noise show that our gradient-correction technique can obtain approximately a 50% reduction in computation time for straight-edge extraction, with a modest improvement in accuracy under most conditions.","PeriodicalId":277395,"journal":{"name":"2009 Digital Image Computing: Techniques and Applications","volume":"107 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116364324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2009 Digital Image Computing: Techniques and Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1