首页 > 最新文献

2010 International Conference on Digital Image Computing: Techniques and Applications最新文献

英文 中文
Differentiating Healthy Cartilage and Damaged Cartilage Using Magnetic Resonance Images in a Quantitative Manner 利用磁共振图像定量鉴别健康软骨和受损软骨
C. Poh, T. K. Chuah, K. Sheah
This paper presents a study that performs a statistical analysis of signal intensities of the cartilage using magnetic resonance images. The aim of the study is to investigate whether it is possible to differentiate cartilage that is normal and cartilage that has damage/lesions in a quantitative manner. Because damaged cartilage tends to have abnormally high signal intensities than that of normal cartilage in fast spin echo proton density weighted (PD) images, we hypothesize that there is a relationship between the signal intensities of the cartilage and the size of the damaged cartilage presence. Twelve MR data sets with different degrees of cartilage damage and five data sets of normal cartilage were used in this study. Femoral articular cartilage was manually segmented using PD images and the MR signal intensities of the cartilage were analyzed. Results show that there is a linear relationship between the difference in mean and median of the cartilage signals (mean-median) and the percentage of damaged cartilage presence (R2 = 0.799, p
本文提出了一项研究,利用磁共振图像对软骨的信号强度进行统计分析。本研究的目的是探讨是否可以定量区分正常软骨和有损伤/病变的软骨。由于在快速自旋回声质子密度加权(PD)图像中,受损软骨往往比正常软骨具有异常高的信号强度,因此我们假设软骨的信号强度与受损软骨的大小之间存在关系。本研究使用了12组不同程度软骨损伤的MR数据集和5组正常软骨的MR数据集。利用PD图像对股骨关节软骨进行人工分割,并对软骨的MR信号强度进行分析。结果表明,软骨信号的均值和中位数之差(均值-中位数)与软骨损伤的存在率呈线性关系(R2 = 0.799, p
{"title":"Differentiating Healthy Cartilage and Damaged Cartilage Using Magnetic Resonance Images in a Quantitative Manner","authors":"C. Poh, T. K. Chuah, K. Sheah","doi":"10.1109/DICTA.2010.98","DOIUrl":"https://doi.org/10.1109/DICTA.2010.98","url":null,"abstract":"This paper presents a study that performs a statistical analysis of signal intensities of the cartilage using magnetic resonance images. The aim of the study is to investigate whether it is possible to differentiate cartilage that is normal and cartilage that has damage/lesions in a quantitative manner. Because damaged cartilage tends to have abnormally high signal intensities than that of normal cartilage in fast spin echo proton density weighted (PD) images, we hypothesize that there is a relationship between the signal intensities of the cartilage and the size of the damaged cartilage presence. Twelve MR data sets with different degrees of cartilage damage and five data sets of normal cartilage were used in this study. Femoral articular cartilage was manually segmented using PD images and the MR signal intensities of the cartilage were analyzed. Results show that there is a linear relationship between the difference in mean and median of the cartilage signals (mean-median) and the percentage of damaged cartilage presence (R2 = 0.799, p","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125882265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CRF Based Region Classification Using Spatial Prototypes 基于CRF的空间原型区域分类
M. Jahangiri, D. Heesch, M. Petrou
This paper proposes a probabilistic model using conditional random field (CRF) for region labelling that encodes and exploits the spatial context of a region. Potential functions for a region depend on a combination of the labels of neighbouring regions as well as their relative location, and a set of typical neighbourhood configurations or prototypes. These are obtained by clustering neighbourhood configurations obtained from a set of annotated images. Inference is achieved by minimising the cost function defined over the CRF model using standard Markov Chain Monte Carlo (MCMC) technique. We validate our approach on a dataset of hand segmented and labelled images of buildings and show that the model outperforms similar such models that utilise either only contextual information or only non-contextual measures.
本文提出了一种利用条件随机场(CRF)进行区域标记的概率模型,该模型对区域的空间上下文进行编码和利用。一个区域的潜在函数依赖于相邻区域的标签及其相对位置的组合,以及一组典型的邻居配置或原型。这些是通过聚类从一组注释图像中获得的邻域配置获得的。通过使用标准马尔可夫链蒙特卡罗(MCMC)技术最小化定义在CRF模型上的成本函数来实现推理。我们在手工分割和标记的建筑物图像数据集上验证了我们的方法,并表明该模型优于仅使用上下文信息或仅使用非上下文度量的类似模型。
{"title":"CRF Based Region Classification Using Spatial Prototypes","authors":"M. Jahangiri, D. Heesch, M. Petrou","doi":"10.1109/DICTA.2010.92","DOIUrl":"https://doi.org/10.1109/DICTA.2010.92","url":null,"abstract":"This paper proposes a probabilistic model using conditional random field (CRF) for region labelling that encodes and exploits the spatial context of a region. Potential functions for a region depend on a combination of the labels of neighbouring regions as well as their relative location, and a set of typical neighbourhood configurations or prototypes. These are obtained by clustering neighbourhood configurations obtained from a set of annotated images. Inference is achieved by minimising the cost function defined over the CRF model using standard Markov Chain Monte Carlo (MCMC) technique. We validate our approach on a dataset of hand segmented and labelled images of buildings and show that the model outperforms similar such models that utilise either only contextual information or only non-contextual measures.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125487665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Robust Extraction of Optic Flow Differentials for Surface Reconstruction 曲面重建中光流微分的鲁棒提取
S. Fu, P. Kovesi
The first-order differential invariants of optic flow, namely divergence, curl, and deformation, provide useful shape indicators of objects passing through view. However, as differential quantities these are often difficult to extract reliably. In this paper we present a filter-based method for computing these invariants with sufficient accuracy to permit the construction of a partial scene model. The noise robustness of our method is analysed using both synthetic and real world images. We also demonstrate that the deformation of a dense optic flow field encodes sufficient information to reliably estimate surface orientations if viewer ego-motion is purely translational.
光流的一阶微分不变量,即散度、旋度和变形,提供了通过视图的物体的有用形状指标。然而,作为微分量,这些通常难以可靠地提取。在本文中,我们提出了一种基于滤波器的方法,以足够的精度计算这些不变量,以允许构建部分场景模型。用合成图像和真实世界图像分析了我们方法的噪声鲁棒性。我们还证明了密集光流场的变形编码了足够的信息来可靠地估计表面方向,如果观察者的自我运动是纯粹的平移。
{"title":"Robust Extraction of Optic Flow Differentials for Surface Reconstruction","authors":"S. Fu, P. Kovesi","doi":"10.1109/DICTA.2010.85","DOIUrl":"https://doi.org/10.1109/DICTA.2010.85","url":null,"abstract":"The first-order differential invariants of optic flow, namely divergence, curl, and deformation, provide useful shape indicators of objects passing through view. However, as differential quantities these are often difficult to extract reliably. In this paper we present a filter-based method for computing these invariants with sufficient accuracy to permit the construction of a partial scene model. The noise robustness of our method is analysed using both synthetic and real world images. We also demonstrate that the deformation of a dense optic flow field encodes sufficient information to reliably estimate surface orientations if viewer ego-motion is purely translational.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130721751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Automatic Recognition of Smiling and Neutral Facial Expressions 微笑和中性面部表情的自动识别
Peiyao Li, S. L. Phung, A. Bouzerdoum, F. Tivive
Facial expression is one way humans convey their emotional states. Accurate recognition of facial expressions via image analysis plays a vital role in perceptual human-computer interaction, robotics and online games. This paper focuses on recognising the smiling from the neutral facial expression. We propose a face alignment method to address the localisation error in existing face detection methods. In this paper, smiling and neutral facial expression are differentiated using a novel neural architecture that combines fixed and adaptive non-linear 2-D filters. The fixed filters are used to extract primitive features, whereas the adaptive filters are trained to extract more complex features for facial expression classification. The proposed approach is evaluated on the JAFFE database and it correctly aligns and crops all images, which is better than several existing methods evaluated on the same database. Our system achieves a classification rate of 99.0% for smiling versus neutral expressions.
面部表情是人类表达情绪状态的一种方式。通过图像分析准确识别面部表情在感知人机交互、机器人和网络游戏中发挥着至关重要的作用。本文主要研究从中性面部表情中识别微笑。针对现有人脸检测方法中存在的定位误差,提出了一种人脸对齐方法。本文使用一种结合固定和自适应非线性二维滤波器的新型神经结构来区分微笑和中性面部表情。固定滤波器用于提取原始特征,而自适应滤波器用于提取更复杂的特征进行面部表情分类。在JAFFE数据库上对该方法进行了评价,结果表明该方法对所有图像进行了正确的对齐和裁剪,优于现有的几种方法。我们的系统对微笑和中性表情的分类率达到了99.0%。
{"title":"Automatic Recognition of Smiling and Neutral Facial Expressions","authors":"Peiyao Li, S. L. Phung, A. Bouzerdoum, F. Tivive","doi":"10.1109/DICTA.2010.103","DOIUrl":"https://doi.org/10.1109/DICTA.2010.103","url":null,"abstract":"Facial expression is one way humans convey their emotional states. Accurate recognition of facial expressions via image analysis plays a vital role in perceptual human-computer interaction, robotics and online games. This paper focuses on recognising the smiling from the neutral facial expression. We propose a face alignment method to address the localisation error in existing face detection methods. In this paper, smiling and neutral facial expression are differentiated using a novel neural architecture that combines fixed and adaptive non-linear 2-D filters. The fixed filters are used to extract primitive features, whereas the adaptive filters are trained to extract more complex features for facial expression classification. The proposed approach is evaluated on the JAFFE database and it correctly aligns and crops all images, which is better than several existing methods evaluated on the same database. Our system achieves a classification rate of 99.0% for smiling versus neutral expressions.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125134973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Graph-Based Text Segmentation Using a Selected Channel Image 使用选定通道图像的基于图形的文本分割
Chao Zeng, W. Jia, Xiangjian He, Jie Yang
This paper proposes a graph-based method for segmentation of a text image using a selected colour-channel image. The text colour information usually presents a two polarity trend. According to the observation that the histogram distributions of the respective colour channel images are usually different from each other, we select the colour channel image with the histogram having the biggest distance between the two main peaks, which represents the main foreground colour strength and background colour strength respectively. The peak distance is estimated by the mean-shift procedure performed on each individual channel image. Then, a graph model is constructed on a selected channel image to segment the text image into foreground and background. The proposed method is tested on a public database, and its effectiveness is demonstrated by the experimental results.
本文提出了一种基于图的方法,使用选定的颜色通道图像对文本图像进行分割。文本色彩信息通常呈现两极性趋势。根据观察到各颜色通道图像的直方图分布通常是不同的,我们选择直方图中两个主峰之间距离最大的颜色通道图像,分别代表主要的前景色彩强度和背景色彩强度。峰值距离通过对每个单独通道图像执行的均值移位程序估计。然后,在选定的通道图像上构建图形模型,将文本图像分割为前景和背景;在一个公共数据库上对该方法进行了测试,实验结果证明了该方法的有效性。
{"title":"Graph-Based Text Segmentation Using a Selected Channel Image","authors":"Chao Zeng, W. Jia, Xiangjian He, Jie Yang","doi":"10.1109/DICTA.2010.95","DOIUrl":"https://doi.org/10.1109/DICTA.2010.95","url":null,"abstract":"This paper proposes a graph-based method for segmentation of a text image using a selected colour-channel image. The text colour information usually presents a two polarity trend. According to the observation that the histogram distributions of the respective colour channel images are usually different from each other, we select the colour channel image with the histogram having the biggest distance between the two main peaks, which represents the main foreground colour strength and background colour strength respectively. The peak distance is estimated by the mean-shift procedure performed on each individual channel image. Then, a graph model is constructed on a selected channel image to segment the text image into foreground and background. The proposed method is tested on a public database, and its effectiveness is demonstrated by the experimental results.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"103 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133762248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Mammographic Mass Detection with Statistical Region Merging 统计区域合并的乳房x线肿块检测
M. Bajger, Fei Ma, Simon Williams, M. Bottema
An automatic method for detection of mammographic masses is presented which utilizes statistical region merging for segmentation (SRM) and linear discriminant analysis (LDA) for classification. The performance of the scheme was evaluated on 36 images selected from the local database of mammograms and on 48 images taken from the Digital Database for Screening Mammography (DDSM). The Az value (area under the ROC curve) for classifying each region was 0.90 for the local dataset and 0.96 for the images from DDSM. Results indicate that SRM segmentation can form part of an robust and efficient basis for analysis of mammograms.
提出了一种利用统计区域合并分割(SRM)和线性判别分析(LDA)进行分类的乳腺肿块自动检测方法。对从本地乳房x线摄影数据库中选择的36张图像和从乳腺x线摄影筛查数字数据库(DDSM)中选择的48张图像进行了性能评估。对每个区域进行分类的Az值(ROC曲线下面积)对于本地数据集为0.90,对于DDSM图像为0.96。结果表明,SRM分割可以构成乳房x线照片分析的稳健和有效基础的一部分。
{"title":"Mammographic Mass Detection with Statistical Region Merging","authors":"M. Bajger, Fei Ma, Simon Williams, M. Bottema","doi":"10.1109/DICTA.2010.14","DOIUrl":"https://doi.org/10.1109/DICTA.2010.14","url":null,"abstract":"An automatic method for detection of mammographic masses is presented which utilizes statistical region merging for segmentation (SRM) and linear discriminant analysis (LDA) for classification. The performance of the scheme was evaluated on 36 images selected from the local database of mammograms and on 48 images taken from the Digital Database for Screening Mammography (DDSM). The Az value (area under the ROC curve) for classifying each region was 0.90 for the local dataset and 0.96 for the images from DDSM. Results indicate that SRM segmentation can form part of an robust and efficient basis for analysis of mammograms.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"199 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134129941","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
3D Curves Reconstruction from Multiple Images 从多个图像重建3D曲线
F. Mai, Y. Hung
In this paper, we propose a new approach for reconstructing 3D curves from a sequence of 2D images taken by uncalibrated cameras. A curve in 3D space is represented by a sequence of 3D points sampled along the curve, and the 3D points are reconstructed by minimizing the distances from their projections to the measured 2D curves on different images (i.e., 2D curve reprojection error). The minimization problem is solved by an iterative algorithm which is guaranteed to converge to a (local) minimum of the 2D reprojection error. Without requiring calibrated cameras or additional point features, our method can reconstruct multiple 3D curves simultaneously from multiple images and it readily handles images with missing and/or partially occluded curves.
在本文中,我们提出了一种新的方法来重建三维曲线从一系列的二维图像由未校准的相机。三维空间中的曲线由沿着曲线采样的一系列三维点表示,通过最小化其投影到不同图像上测量的二维曲线的距离(即二维曲线重投影误差)来重建三维点。最小化问题采用迭代算法求解,保证收敛到二维重投影误差的(局部)最小值。不需要校准相机或额外的点特征,我们的方法可以从多个图像同时重建多个3D曲线,并且它很容易处理丢失和/或部分遮挡曲线的图像。
{"title":"3D Curves Reconstruction from Multiple Images","authors":"F. Mai, Y. Hung","doi":"10.1109/DICTA.2010.84","DOIUrl":"https://doi.org/10.1109/DICTA.2010.84","url":null,"abstract":"In this paper, we propose a new approach for reconstructing 3D curves from a sequence of 2D images taken by uncalibrated cameras. A curve in 3D space is represented by a sequence of 3D points sampled along the curve, and the 3D points are reconstructed by minimizing the distances from their projections to the measured 2D curves on different images (i.e., 2D curve reprojection error). The minimization problem is solved by an iterative algorithm which is guaranteed to converge to a (local) minimum of the 2D reprojection error. Without requiring calibrated cameras or additional point features, our method can reconstruct multiple 3D curves simultaneously from multiple images and it readily handles images with missing and/or partially occluded curves.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129292498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
An Increased Throughput FPGA Design of the JPEG2000 Binary Arithmetic Decoder JPEG2000二进制算术解码器的高吞吐量FPGA设计
D. Lucking, E. Balster
As digital imaging techniques continue to advance, new image compression standards are needed to keep the transmission time and storage space low for increasing image sizes. The Joint Photographic Expert Group (JPEG) fulfilled this need with the ratification of the JPEG2000 standard in December of 2000. JPEG2000 adds many features to image compression technology but also increases the computational complexity of traditional encoders. To mitigate the added computational complexity, the JPEG2000 algorithm allows processing parts in parallel, increasing the benefits of implementing the algorithm in application specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs). A ¿exible FPGA implementation of the JPEG2000 binary arithmetic decoder, the core component of the JPEG2000 decoding algorithm, is presented in this paper. The proposed JPEG2000 binary arithmetic decoder reduces the amount of resources used on the FPGA allowing 17% more entropy block decoders to fit on chip and consequently increasing the throughput by 35% beyond previous designs.
随着数字成像技术的不断发展,需要新的图像压缩标准来保持传输时间和存储空间低,以增加图像尺寸。联合摄影专家组(JPEG)在2000年12月批准了JPEG2000标准,满足了这一需求。JPEG2000为图像压缩技术增加了许多特性,但也增加了传统编码器的计算复杂度。为了减轻增加的计算复杂性,JPEG2000算法允许并行处理部件,增加了在特定应用集成电路(asic)或现场可编程门阵列(fpga)中实现算法的好处。本文介绍了JPEG2000二进制算术解码器的FPGA实现方法,该解码器是JPEG2000译码算法的核心部件。提出的JPEG2000二进制算法解码器减少了FPGA上使用的资源量,允许在芯片上安装17%的熵块解码器,从而比以前的设计提高了35%的吞吐量。
{"title":"An Increased Throughput FPGA Design of the JPEG2000 Binary Arithmetic Decoder","authors":"D. Lucking, E. Balster","doi":"10.1109/DICTA.2010.74","DOIUrl":"https://doi.org/10.1109/DICTA.2010.74","url":null,"abstract":"As digital imaging techniques continue to advance, new image compression standards are needed to keep the transmission time and storage space low for increasing image sizes. The Joint Photographic Expert Group (JPEG) fulfilled this need with the ratification of the JPEG2000 standard in December of 2000. JPEG2000 adds many features to image compression technology but also increases the computational complexity of traditional encoders. To mitigate the added computational complexity, the JPEG2000 algorithm allows processing parts in parallel, increasing the benefits of implementing the algorithm in application specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs). A ¿exible FPGA implementation of the JPEG2000 binary arithmetic decoder, the core component of the JPEG2000 decoding algorithm, is presented in this paper. The proposed JPEG2000 binary arithmetic decoder reduces the amount of resources used on the FPGA allowing 17% more entropy block decoders to fit on chip and consequently increasing the throughput by 35% beyond previous designs.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"188 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133858398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Registration Invariant Representations for Expression Detection 表达式检测的注册不变表示
P. Lucey, S. Lucey, J. Cohn
Active appearance model (AAM) representations have been used to great effect recently in the accurate detection of expression events (e.g., action units, pain, broad expressions, etc.). The motivation for their use, and rationale for their success, lies in their ability to: (i) provide dense (i.e. 60- 70 points on the face) registration accuracy on par with a human labeler, and (ii) the ability to decompose the registered face image to separate appearance and shape representations. Unfortunately, this human-like registration performance is isolated to registration algorithms that are specifically tuned to the illumination, camera and subject being tracked (i.e. "subject dependent'' algorithms). As a result, it is rare, to see AAM representations being employed in the far more useful "subject independent'' situations (i.e., where illumination, camera and subject is unknown) due to the inherent increased geometric noise present in the estimated registration. In this paper we argue that "AAM like'' expression detection results can be obtained in the presence of noisy dense registration through the employment of registration invariant representations (e.g., Gabor magnitudes and HOG features). We demonstrate that good expression detection performance can still be enjoyed over the types of geometric noise often encountered with the more geometrically noisy state of the art generic algorithms (e.g., Bayesian Tangent Shape Models (BTSM), Constrained Local Models (CLM), etc). We show these results on the extended Cohn-Kanade (CK+) database over all facial action units.
近年来,活动外观模型(AAM)表征在表情事件(如动作单元、疼痛、广义表情等)的准确检测中发挥了重要作用。使用它们的动机和它们成功的基本原理在于它们能够:(i)提供与人类标记器相当的密集(即面部上的60- 70个点)注册精度,以及(ii)分解注册的人脸图像以分离外观和形状表示的能力。不幸的是,这种类似人类的注册性能与专门针对照明、相机和被跟踪对象进行调整的注册算法是隔离的。“主题相关”算法)。因此,由于估计配准中存在固有的几何噪声增加,因此很少看到AAM表示被用于更有用的“主体独立”情况(即,照明,相机和主体未知的情况)。在本文中,我们认为通过使用配准不变量表示(例如,Gabor幅度和HOG特征),可以在存在噪声密集配准的情况下获得“AAM样”表达检测结果。我们证明了良好的表达检测性能仍然可以享受到更多几何噪声的最先进的通用算法(例如,贝叶斯切线形状模型(BTSM),约束局部模型(CLM)等)经常遇到的几何噪声类型。我们在所有面部动作单元的扩展Cohn-Kanade (CK+)数据库上展示了这些结果。
{"title":"Registration Invariant Representations for Expression Detection","authors":"P. Lucey, S. Lucey, J. Cohn","doi":"10.1109/DICTA.2010.53","DOIUrl":"https://doi.org/10.1109/DICTA.2010.53","url":null,"abstract":"Active appearance model (AAM) representations have been used to great effect recently in the accurate detection of expression events (e.g., action units, pain, broad expressions, etc.). The motivation for their use, and rationale for their success, lies in their ability to: (i) provide dense (i.e. 60- 70 points on the face) registration accuracy on par with a human labeler, and (ii) the ability to decompose the registered face image to separate appearance and shape representations. Unfortunately, this human-like registration performance is isolated to registration algorithms that are specifically tuned to the illumination, camera and subject being tracked (i.e. \"subject dependent'' algorithms). As a result, it is rare, to see AAM representations being employed in the far more useful \"subject independent'' situations (i.e., where illumination, camera and subject is unknown) due to the inherent increased geometric noise present in the estimated registration. In this paper we argue that \"AAM like'' expression detection results can be obtained in the presence of noisy dense registration through the employment of registration invariant representations (e.g., Gabor magnitudes and HOG features). We demonstrate that good expression detection performance can still be enjoyed over the types of geometric noise often encountered with the more geometrically noisy state of the art generic algorithms (e.g., Bayesian Tangent Shape Models (BTSM), Constrained Local Models (CLM), etc). We show these results on the extended Cohn-Kanade (CK+) database over all facial action units.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"53-54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131004538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Appearance-Based Re-identification of People in Video 基于外貌的视频人物再识别
Arif Khan, Jian Zhang, Yang Wang
This paper introduces the topic of appearance-based re-identification of people in video. This work is based on colour information of people’s clothing. Most of the work described in the literature uses full body histogram. This paper evaluates the histogram method and describes ways of including spatial colour information. The paper proposes a colour-based appearance descriptor called Colour Context People Descriptor. All the methods are evaluated extensively. The results are reported in the experiments. It is concluded at the end that adding spatial colour information greatly improves the re-identification results.
本文介绍了视频中基于外貌的人物再识别问题。这个作品是基于人们服装的颜色信息。文献中描述的大部分工作使用全身直方图。本文对直方图方法进行了评价,并描述了包含空间颜色信息的方法。本文提出了一种基于颜色的外观描述符,称为颜色上下文人物描述符。所有的方法都进行了广泛的评估。在实验中报告了结果。最后得出结论,加入空间颜色信息可以大大改善再识别结果。
{"title":"Appearance-Based Re-identification of People in Video","authors":"Arif Khan, Jian Zhang, Yang Wang","doi":"10.1109/DICTA.2010.67","DOIUrl":"https://doi.org/10.1109/DICTA.2010.67","url":null,"abstract":"This paper introduces the topic of appearance-based re-identification of people in video. This work is based on colour information of people’s clothing. Most of the work described in the literature uses full body histogram. This paper evaluates the histogram method and describes ways of including spatial colour information. The paper proposes a colour-based appearance descriptor called Colour Context People Descriptor. All the methods are evaluated extensively. The results are reported in the experiments. It is concluded at the end that adding spatial colour information greatly improves the re-identification results.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130507403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
期刊
2010 International Conference on Digital Image Computing: Techniques and Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1