首页 > 最新文献

2010 International Conference on Digital Image Computing: Techniques and Applications最新文献

英文 中文
Harvesting Web Images for Realistic Facial Expression Recognition 采集网络图像的现实面部表情识别
Kaimin Yu, Zhiyong Wang, L. Zhuo, D. Feng
Large amount of labeled training data is required to develop robust and effective facial expression analysis methods. However, obtaining such data is typically a tedious and time-consuming task that is proportional to the size of the database. Due to the rapid advance of Internet and Web technologies, it is now feasible to collect a tremendous number of images with potential label information at a low cost of human effort. Therefore, this paper proposes a framework to collect realistic facial expression images from the web so as to promote further research on robust facial expression recognition. Due to the limitation of current commercial web search engines, a large fraction of returned images is not related to the query keyword. We present a SVM based active learning approach to selecting relevant images from noisy image search results. The resulting database is more diverse with more sample images, compared with other well established facial expression databases CK and JAFFE. Experimental results demonstrate that the generalization of our web based database outperforms those two existing databases. It is anticipated that further research on facial expression recognition or even affective computing will not be restricted to traditional 7 categories only.
开发鲁棒有效的面部表情分析方法需要大量的标记训练数据。然而,获取此类数据通常是一项繁琐且耗时的任务,与数据库的大小成正比。由于Internet和Web技术的快速发展,现在可以以较低的人力成本收集大量具有潜在标签信息的图像。因此,本文提出了一个从网络中收集真实面部表情图像的框架,以促进鲁棒性面部表情识别的进一步研究。由于目前商业网络搜索引擎的限制,很大一部分返回的图像与查询关键字无关。提出了一种基于支持向量机的主动学习方法,从噪声图像搜索结果中选择相关图像。与其他成熟的面部表情数据库CK和JAFFE相比,由此产生的数据库具有更多的样本图像,更加多样化。实验结果表明,基于web的数据库的泛化性能优于现有的两种数据库。预计面部表情识别甚至情感计算的进一步研究将不仅仅局限于传统的7个类别。
{"title":"Harvesting Web Images for Realistic Facial Expression Recognition","authors":"Kaimin Yu, Zhiyong Wang, L. Zhuo, D. Feng","doi":"10.1109/DICTA.2010.93","DOIUrl":"https://doi.org/10.1109/DICTA.2010.93","url":null,"abstract":"Large amount of labeled training data is required to develop robust and effective facial expression analysis methods. However, obtaining such data is typically a tedious and time-consuming task that is proportional to the size of the database. Due to the rapid advance of Internet and Web technologies, it is now feasible to collect a tremendous number of images with potential label information at a low cost of human effort. Therefore, this paper proposes a framework to collect realistic facial expression images from the web so as to promote further research on robust facial expression recognition. Due to the limitation of current commercial web search engines, a large fraction of returned images is not related to the query keyword. We present a SVM based active learning approach to selecting relevant images from noisy image search results. The resulting database is more diverse with more sample images, compared with other well established facial expression databases CK and JAFFE. Experimental results demonstrate that the generalization of our web based database outperforms those two existing databases. It is anticipated that further research on facial expression recognition or even affective computing will not be restricted to traditional 7 categories only.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121693890","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Adaptive Stick-Like Features for Human Detection Based on Multi-scale Feature Fusion Scheme 基于多尺度特征融合方案的自适应棒状特征人体检测
Sheng Wang, Ruo Du, Qiang Wu, Xiangjian He
Human detection has been widely used in many applications. In the meantime, it is still a difficult problem with many open questions due to challenges caused by various factors such as clothing, posture and etc. By investigating several benchmark methods and frameworks in the literature, this paper proposes a novel method which successfully implements the Real AdaBoost training procedure on multi-scale images. Various object features are exposed on multiple levels. To further boost the overall performance, a fusion scheme is established using scores obtained at various levels which integrates decision results with different scales to make the final decision. Unlike other score-based fusion methods, this paper re-formulates the fusion process through a supervised learning. Therefore, our fusion approach can better distinguish subtle difference between human objects and non-human objects. Furthermore, in our approach, we are able to use simpler weak features for boosting and hence alleviate the training complexity existed in most of AdaBoost training approaches. Encouraging results are obtained on a well recognized benchmark database.
人体检测已广泛应用于许多领域。同时,由于服装、姿势等各种因素带来的挑战,这仍然是一个难题,有许多悬而未决的问题。通过对文献中几种基准方法和框架的研究,本文提出了一种新的方法,该方法成功地在多尺度图像上实现了Real AdaBoost训练过程。在多个层次上暴露各种对象特征。为了进一步提高整体性能,建立了一种融合方案,将不同尺度的决策结果进行融合,从而做出最终决策。与其他基于分数的融合方法不同,本文通过监督学习重新制定融合过程。因此,我们的融合方法可以更好地区分人类物体和非人类物体之间的细微差异。此外,在我们的方法中,我们能够使用更简单的弱特征进行增强,从而减轻了大多数AdaBoost训练方法中存在的训练复杂性。在一个公认的基准数据库上获得了令人鼓舞的结果。
{"title":"Adaptive Stick-Like Features for Human Detection Based on Multi-scale Feature Fusion Scheme","authors":"Sheng Wang, Ruo Du, Qiang Wu, Xiangjian He","doi":"10.1109/DICTA.2010.70","DOIUrl":"https://doi.org/10.1109/DICTA.2010.70","url":null,"abstract":"Human detection has been widely used in many applications. In the meantime, it is still a difficult problem with many open questions due to challenges caused by various factors such as clothing, posture and etc. By investigating several benchmark methods and frameworks in the literature, this paper proposes a novel method which successfully implements the Real AdaBoost training procedure on multi-scale images. Various object features are exposed on multiple levels. To further boost the overall performance, a fusion scheme is established using scores obtained at various levels which integrates decision results with different scales to make the final decision. Unlike other score-based fusion methods, this paper re-formulates the fusion process through a supervised learning. Therefore, our fusion approach can better distinguish subtle difference between human objects and non-human objects. Furthermore, in our approach, we are able to use simpler weak features for boosting and hence alleviate the training complexity existed in most of AdaBoost training approaches. Encouraging results are obtained on a well recognized benchmark database.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"132 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115967849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Human Action Recognition from Boosted Pose Estimation 基于增强姿态估计的人体动作识别
Li Wang, Li Cheng, Tuan Hue Thi, Jian Zhang
This paper presents a unified framework for recognizing human action in video using human pose estimation. Due to high variation of human appearance and noisy context background, accurate human pose analysis is hard to achieve and rarely employed for the task of action recognition. In our approach, we take advantage of the current success of human detection and view invariability of local feature-based approach to design a pose-based action recognition system. We begin with a frame-wise human detection step to initialize the search space for human local parts, then integrate the detected parts into human kinematic structure using a tree structural graphical model. The final human articulation configuration is eventually used to infer the action class being performed based on each single part behavior and the overall structure variation. In our work, we also show that even with imprecise pose estimation, accurate action recognition can still be achieved based on informative clues from the overall pose part configuration. The promising results obtained from action recognition benchmark have proven our proposed framework is comparable to the existing state-of-the-art action recognition algorithms.
本文提出了一种基于人体姿态估计的视频中人体动作识别的统一框架。由于人体外观的高度变化和嘈杂的环境背景,很难实现准确的人体姿势分析,很少用于动作识别任务。在我们的方法中,我们利用当前成功的基于局部特征的人类检测和视图不变性的方法来设计一个基于姿态的动作识别系统。我们从逐帧的人体检测步骤开始,初始化人体局部部位的搜索空间,然后使用树结构图形模型将检测到的部位整合到人体运动结构中。最终的人体关节配置最终用于推断基于每个单个部分行为和整体结构变化所执行的动作类。在我们的工作中,我们还表明,即使不精确的姿态估计,仍然可以基于来自整体姿态部分配置的信息线索实现准确的动作识别。从动作识别基准测试中获得的结果表明,我们提出的框架与现有的最先进的动作识别算法相当。
{"title":"Human Action Recognition from Boosted Pose Estimation","authors":"Li Wang, Li Cheng, Tuan Hue Thi, Jian Zhang","doi":"10.1109/DICTA.2010.60","DOIUrl":"https://doi.org/10.1109/DICTA.2010.60","url":null,"abstract":"This paper presents a unified framework for recognizing human action in video using human pose estimation. Due to high variation of human appearance and noisy context background, accurate human pose analysis is hard to achieve and rarely employed for the task of action recognition. In our approach, we take advantage of the current success of human detection and view invariability of local feature-based approach to design a pose-based action recognition system. We begin with a frame-wise human detection step to initialize the search space for human local parts, then integrate the detected parts into human kinematic structure using a tree structural graphical model. The final human articulation configuration is eventually used to infer the action class being performed based on each single part behavior and the overall structure variation. In our work, we also show that even with imprecise pose estimation, accurate action recognition can still be achieved based on informative clues from the overall pose part configuration. The promising results obtained from action recognition benchmark have proven our proposed framework is comparable to the existing state-of-the-art action recognition algorithms.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123819106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Robust Dimensionality Reduction for Human Action Recognition 基于鲁棒降维的人体动作识别
Óscar Pérez, R. Xu, M. Piccardi
Human action recognition can be approached by combining an action-discriminative feature set with a classifier. However, the dimensionality of typical feature sets joint with that of the time dimension often leads to a curse-of-dimensionality situation. Moreover, the measurement of the feature set is subject to sometime severe errors. This paper presents an approach to human action recognition based on robust dimensionality reduction. The observation probabilities of hidden Markov models (HMM) are modelled by mixtures of probabilistic principal components analyzers and mixtures of $t$-distribution sub-spaces, and compared with conventional Gaussian mixture models. Experimental results on two datasets show that dimensionality reduction helps improve the classification accuracy and that the heavier-tailed $t$-distribution can help reduce the impact of outliers generated by segmentation errors.
人类动作识别可以通过将动作判别特征集与分类器相结合来实现。然而,典型特征集的维数与时间维数相结合时,往往会出现维数不足的情况。此外,特征集的测量有时会出现严重的错误。提出了一种基于鲁棒降维的人体动作识别方法。将隐马尔可夫模型(HMM)的观测概率用概率主成分分析器和t -分布子空间的混合模型来建模,并与传统的高斯混合模型进行了比较。在两个数据集上的实验结果表明,降维有助于提高分类精度,重尾的$t$-分布有助于减少分割错误产生的离群值的影响。
{"title":"Robust Dimensionality Reduction for Human Action Recognition","authors":"Óscar Pérez, R. Xu, M. Piccardi","doi":"10.1109/DICTA.2010.66","DOIUrl":"https://doi.org/10.1109/DICTA.2010.66","url":null,"abstract":"Human action recognition can be approached by combining an action-discriminative feature set with a classifier. However, the dimensionality of typical feature sets joint with that of the time dimension often leads to a curse-of-dimensionality situation. Moreover, the measurement of the feature set is subject to sometime severe errors. This paper presents an approach to human action recognition based on robust dimensionality reduction. The observation probabilities of hidden Markov models (HMM) are modelled by mixtures of probabilistic principal components analyzers and mixtures of $t$-distribution sub-spaces, and compared with conventional Gaussian mixture models. Experimental results on two datasets show that dimensionality reduction helps improve the classification accuracy and that the heavier-tailed $t$-distribution can help reduce the impact of outliers generated by segmentation errors.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121533582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Multiple Homography Estimation with Full Consistency Constraints 具有完全一致性约束的多重单应性估计
W. Chojnacki, Zygmunt L. Szpak, M. Brooks, A. Hengel
A novel approach is presented to estimating a set of interdependent homography matrices linked together by latent variables. The approach allows enforcement of all underlying consistency constraints while accounting for the arbitrariness of the scale of each individual matrix. The input data is assumed to be in the form of a set of homography matrices obtained by estimation from image data with the consistency constraints ignored, appended by a set of error covariances associated with these matrix estimates. A cost function is proposed for upgrading, via optimisation, the input data to a set of homography matrices satisfying the constraints. The function is invariant to a change of any of the individual scales of the input matrices. The proposed approach is applied to the particular problem of estimating a set of homography matrices induced by multiple planes in the 3D scene between two views. Experimental results are given which demonstrate the effectiveness of the approach.
提出了一种新的方法来估计一组由潜在变量连接在一起的相互依赖的单应矩阵。该方法允许执行所有潜在的一致性约束,同时考虑到每个单独矩阵的规模的任意性。假设输入数据是由忽略一致性约束的图像数据估计得到的一组单应性矩阵的形式,并附加一组与这些矩阵估计相关的误差协方差。提出了一个成本函数,通过优化,将输入数据升级为满足约束的一组单应矩阵。该函数对于输入矩阵的任何单个尺度的变化都是不变的。该方法应用于三维场景中两个视图之间由多个平面引起的一组单应矩阵的估计问题。实验结果表明了该方法的有效性。
{"title":"Multiple Homography Estimation with Full Consistency Constraints","authors":"W. Chojnacki, Zygmunt L. Szpak, M. Brooks, A. Hengel","doi":"10.1109/DICTA.2010.87","DOIUrl":"https://doi.org/10.1109/DICTA.2010.87","url":null,"abstract":"A novel approach is presented to estimating a set of interdependent homography matrices linked together by latent variables. The approach allows enforcement of all underlying consistency constraints while accounting for the arbitrariness of the scale of each individual matrix. The input data is assumed to be in the form of a set of homography matrices obtained by estimation from image data with the consistency constraints ignored, appended by a set of error covariances associated with these matrix estimates. A cost function is proposed for upgrading, via optimisation, the input data to a set of homography matrices satisfying the constraints. The function is invariant to a change of any of the individual scales of the input matrices. The proposed approach is applied to the particular problem of estimating a set of homography matrices induced by multiple planes in the 3D scene between two views. Experimental results are given which demonstrate the effectiveness of the approach.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127535889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Unsupervised Detection for Minimizing a Region of Interest around Distinct Object in Natural Images 自然图像中不同目标周围兴趣区域最小化的无监督检测
Anucha Tungkatsathan, W. Premchaiswadi, Nucharee Premchaiswadi
One of the major challenges for region-based image retrieval is to identify the Region of Interest (ROI) that comprises object queries. However, automatically identifying the regions or objects of interest in a natural scene is a very difficult task because the content is complex and can be any shape. In this paper, we present a novel unsupervised detection method to automatically and efficiently minimize the ROI in the images. We applied an edge-based active contour model that drew upon edge information in local regions. The mathematical implementation of the proposed active contour model was accomplished using a variational level set formulation. In addition, the mean-shift algorithm was used to reduce the sensitivity of parameter change of level set formulation. The results show that our method can overcome the difficulties of non-uniform sub-region and intensity in homogeneities in natural image segmentation.
基于区域的图像检索面临的主要挑战之一是识别包含对象查询的感兴趣区域(ROI)。然而,在自然场景中自动识别感兴趣的区域或物体是一项非常困难的任务,因为内容很复杂,可以是任何形状。在本文中,我们提出了一种新的无监督检测方法来自动有效地最小化图像中的ROI。我们应用了一种基于边缘的活动轮廓模型,该模型利用了局部区域的边缘信息。利用变分水平集公式实现了所提出的活动轮廓模型的数学实现。此外,采用mean-shift算法降低了水平集公式参数变化的敏感性。结果表明,该方法克服了自然图像分割中均匀性中子区域和强度不均匀的困难。
{"title":"Unsupervised Detection for Minimizing a Region of Interest around Distinct Object in Natural Images","authors":"Anucha Tungkatsathan, W. Premchaiswadi, Nucharee Premchaiswadi","doi":"10.1109/DICTA.2010.45","DOIUrl":"https://doi.org/10.1109/DICTA.2010.45","url":null,"abstract":"One of the major challenges for region-based image retrieval is to identify the Region of Interest (ROI) that comprises object queries. However, automatically identifying the regions or objects of interest in a natural scene is a very difficult task because the content is complex and can be any shape. In this paper, we present a novel unsupervised detection method to automatically and efficiently minimize the ROI in the images. We applied an edge-based active contour model that drew upon edge information in local regions. The mathematical implementation of the proposed active contour model was accomplished using a variational level set formulation. In addition, the mean-shift algorithm was used to reduce the sensitivity of parameter change of level set formulation. The results show that our method can overcome the difficulties of non-uniform sub-region and intensity in homogeneities in natural image segmentation.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"2005 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128809448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Focusing the Normalised Information Distance on the Relevant Information Content for Image Similarity 将归一化信息距离聚焦在相关信息内容上实现图像相似性
Joselíto J. Chua, P. Tischer
This paper investigates the normalised information distance (NID) proposed by Bennet et~al~(1998) as an approach to measure the visual similarity (or dissimilarity) of images. Earlier studies suggest that compression-based approximations to the NID can yield dissimilarity measures that correlate well with visual comparisons. However, results also indicate that conventional feature-based dissimilarity measures often outperform those that are based on the NID. This paper proposes that a theoretical decomposition of the NID can help explain why the NID-based dissimilarity measures might not perform well compared to feature-based approaches. The theoretical decomposition considers the perceptually relevant and irrelevant information content for image similarity. We illustrate how the NID-based dissimilarity measures could be improved by discarding the irrelevant information, and applying the NID on only the relevant information.
本文研究了bennett等人(1998)提出的归一化信息距离(NID)作为测量图像视觉相似性(或不相似性)的方法。早期的研究表明,基于压缩的NID近似可以产生与视觉比较相关的不相似性测量。然而,结果也表明,传统的基于特征的不相似度度量通常优于基于NID的不相似度度量。本文提出,NID的理论分解可以帮助解释为什么基于NID的不相似性度量与基于特征的方法相比可能表现不佳。理论分解考虑了图像相似性的感知相关和不相关信息内容。我们说明了如何通过丢弃不相关信息并仅对相关信息应用NID来改进基于NID的不相似度度量。
{"title":"Focusing the Normalised Information Distance on the Relevant Information Content for Image Similarity","authors":"Joselíto J. Chua, P. Tischer","doi":"10.1109/DICTA.2010.10","DOIUrl":"https://doi.org/10.1109/DICTA.2010.10","url":null,"abstract":"This paper investigates the normalised information distance (NID) proposed by Bennet et~al~(1998) as an approach to measure the visual similarity (or dissimilarity) of images. Earlier studies suggest that compression-based approximations to the NID can yield dissimilarity measures that correlate well with visual comparisons. However, results also indicate that conventional feature-based dissimilarity measures often outperform those that are based on the NID. This paper proposes that a theoretical decomposition of the NID can help explain why the NID-based dissimilarity measures might not perform well compared to feature-based approaches. The theoretical decomposition considers the perceptually relevant and irrelevant information content for image similarity. We illustrate how the NID-based dissimilarity measures could be improved by discarding the irrelevant information, and applying the NID on only the relevant information.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132441246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
L1/2 Sparsity Constrained Nonnegative Matrix Factorization for Hyperspectral Unmixing 高光谱分解的L1/2稀疏约束非负矩阵分解
Y. Qian, Sen Jia, J. Zhou, A. Robles-Kelly
Hyperspectral unmixing is a crucial preprocessing step for material classification and recognition. In the last decade, nonnegative matrix factorization (NMF) and its extensions have been intensively studied to unmix hyperspectral imagery and recover the material end-members. As an important constraint, sparsity has been modeled making use of L1 or L2 regularizers. However, the full additivity constraint of material abundances is often overlooked, hence, limiting the practical efficacy of these methods. In this paper, we extend the NMF algorithm by incorporating the L1/2 sparsity constraint. The L1/2-NMF provides more sparse and accurate results than the other regularizers by considering the end-member additivity constraint explicitly in the optimisation process. Experiments on the synthetic and real hyperspectral data validate the proposed algorithm.
高光谱解调是材料分类识别的关键预处理步骤。近十年来,人们对非负矩阵分解(NMF)及其扩展进行了深入的研究,以解混高光谱图像并恢复材料端元。作为一个重要的约束,稀疏性已经利用L1或L2正则化器建模。然而,材料丰度的全加性约束往往被忽视,因此,限制了这些方法的实际效果。本文通过引入L1/2稀疏性约束对NMF算法进行了扩展。L1/2-NMF通过在优化过程中明确考虑端元可加性约束,提供了比其他正则化器更稀疏和准确的结果。在合成高光谱数据和真实高光谱数据上的实验验证了该算法的有效性。
{"title":"L1/2 Sparsity Constrained Nonnegative Matrix Factorization for Hyperspectral Unmixing","authors":"Y. Qian, Sen Jia, J. Zhou, A. Robles-Kelly","doi":"10.1109/DICTA.2010.82","DOIUrl":"https://doi.org/10.1109/DICTA.2010.82","url":null,"abstract":"Hyperspectral unmixing is a crucial preprocessing step for material classification and recognition. In the last decade, nonnegative matrix factorization (NMF) and its extensions have been intensively studied to unmix hyperspectral imagery and recover the material end-members. As an important constraint, sparsity has been modeled making use of L1 or L2 regularizers. However, the full additivity constraint of material abundances is often overlooked, hence, limiting the practical efficacy of these methods. In this paper, we extend the NMF algorithm by incorporating the L1/2 sparsity constraint. The L1/2-NMF provides more sparse and accurate results than the other regularizers by considering the end-member additivity constraint explicitly in the optimisation process. Experiments on the synthetic and real hyperspectral data validate the proposed algorithm.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132259929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
IMU-Aided 3D Reconstruction Based on Multiple Virtual Planes 基于多虚拟平面的imu辅助三维重建
H. Aliakbarpour, J. Dias
This paper proposes a novel approach for fast 3D reconstruction of an object inside a scene by using Inertial Measurement Unit (IMU) data. A network of cameras is used to observe the scene. For each camera within the network, a virtual camera is considered by using the concept of emph{infinite homography}. Such a virtual camera is downward and has optical axis parallel to the gravity vector. Then a set of virtual horizontal 3D planes are considered for the aim of 3D reconstruction. The intersection of these virtual parallel 3D planes with the object is computed using the concept of homography and by applying a 2D Bayesian occupancy grid for each plane. The experimental results validate both feasibility and effectiveness of the proposed method.
本文提出了一种利用惯性测量单元(IMU)数据实现场景内物体快速三维重建的新方法。一个摄像机网络被用来观察现场。对于网络中的每个摄像机,使用emph{无限单应性}的概念来考虑一个虚拟摄像机。这种虚拟相机是向下的,其光轴平行于重力矢量。然后考虑了一组虚拟的三维水平平面来实现三维重建。这些虚拟的平行三维平面与物体的交点使用单应性的概念,并通过对每个平面应用二维贝叶斯占用网格来计算。实验结果验证了该方法的可行性和有效性。
{"title":"IMU-Aided 3D Reconstruction Based on Multiple Virtual Planes","authors":"H. Aliakbarpour, J. Dias","doi":"10.1109/DICTA.2010.86","DOIUrl":"https://doi.org/10.1109/DICTA.2010.86","url":null,"abstract":"This paper proposes a novel approach for fast 3D reconstruction of an object inside a scene by using Inertial Measurement Unit (IMU) data. A network of cameras is used to observe the scene. For each camera within the network, a virtual camera is considered by using the concept of emph{infinite homography}. Such a virtual camera is downward and has optical axis parallel to the gravity vector. Then a set of virtual horizontal 3D planes are considered for the aim of 3D reconstruction. The intersection of these virtual parallel 3D planes with the object is computed using the concept of homography and by applying a 2D Bayesian occupancy grid for each plane. The experimental results validate both feasibility and effectiveness of the proposed method.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124197307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Improved Reconstruction of Flutter Shutter Images for Motion Blur Reduction 改进的颤振快门图像重建运动模糊减少
A. Sarker, Len Hamey
Relative motion between a camera and its subject introduces motion blur in captured images. Reconstruction of unblurred images is ill-posed due to the loss of spatial high frequencies. The flutter shutter preserves high frequencies by rapidly opening and closing the shutter during exposure, providing greatly improved reconstruction. We address two open problems in the reconstruction of unblurred images from flutter shutter images. Firstly, we propose a noise reduction technique that reduces reconstruction noise while preserving image detail. Secondly, we propose a semi-automatic technique for estimating the Point Spread Function of the motion blur. Together these techniques provide substantial improvement in reconstruction of flutter shutter images.
相机与其拍摄对象之间的相对运动会在拍摄的图像中引入运动模糊。由于空间高频的损失,未模糊图像的重建是病态的。颤振快门通过在曝光过程中快速打开和关闭快门来保持高频率,从而大大改善了重建。我们解决了从颤振快门图像中重建未模糊图像的两个开放问题。首先,我们提出了一种在保留图像细节的同时降低重构噪声的降噪技术。其次,提出了一种半自动估计运动模糊点扩散函数的方法。这些技术共同为颤振快门图像的重建提供了实质性的改进。
{"title":"Improved Reconstruction of Flutter Shutter Images for Motion Blur Reduction","authors":"A. Sarker, Len Hamey","doi":"10.1109/DICTA.2010.77","DOIUrl":"https://doi.org/10.1109/DICTA.2010.77","url":null,"abstract":"Relative motion between a camera and its subject introduces motion blur in captured images. Reconstruction of unblurred images is ill-posed due to the loss of spatial high frequencies. The flutter shutter preserves high frequencies by rapidly opening and closing the shutter during exposure, providing greatly improved reconstruction. We address two open problems in the reconstruction of unblurred images from flutter shutter images. Firstly, we propose a noise reduction technique that reduces reconstruction noise while preserving image detail. Secondly, we propose a semi-automatic technique for estimating the Point Spread Function of the motion blur. Together these techniques provide substantial improvement in reconstruction of flutter shutter images.","PeriodicalId":246460,"journal":{"name":"2010 International Conference on Digital Image Computing: Techniques and Applications","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121187278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
2010 International Conference on Digital Image Computing: Techniques and Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1