首页 > 最新文献

2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops最新文献

英文 中文
Computing iconic summaries of general visual concepts 计算一般视觉概念的图标摘要
R. Raguram, S. Lazebnik
This paper considers the problem of selecting iconic images to summarize general visual categories. We define iconic images as high-quality representatives of a large group of images consistent both in appearance and semantics. To find such groups, we perform joint clustering in the space of global image descriptors and latent topic vectors of tags associated with the images. To select the representative iconic images for the joint clusters, we use a quality ranking learned from a large collection of labeled images. For the purposes of visualization, iconic images are grouped by semantic ldquothemerdquo and multidimensional scaling is used to compute a 2D layout that reflects the relationships between the themes. Results on four large-scale datasets demonstrate the ability of our approach to discover plausible themes and recurring visual motifs for challenging abstract concepts such as ldquoloverdquo and ldquobeautyrdquo.
本文考虑了选取标志性图像来总结一般视觉类别的问题。我们将标志性图像定义为在外观和语义上一致的一大组图像的高质量代表。为了找到这样的组,我们在全局图像描述符和与图像相关的标签的潜在主题向量的空间中进行联合聚类。为了为联合聚类选择代表性的标志性图像,我们使用了从大量标记图像中学习到的质量排名。为了可视化的目的,图标图像按语义分组,并使用多维缩放来计算反映主题之间关系的2D布局。在四个大规模数据集上的结果表明,我们的方法能够为挑战抽象概念(如ldquoloverdquo和ldquobeautydquo)发现合理的主题和反复出现的视觉主题。
{"title":"Computing iconic summaries of general visual concepts","authors":"R. Raguram, S. Lazebnik","doi":"10.1109/CVPRW.2008.4562959","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562959","url":null,"abstract":"This paper considers the problem of selecting iconic images to summarize general visual categories. We define iconic images as high-quality representatives of a large group of images consistent both in appearance and semantics. To find such groups, we perform joint clustering in the space of global image descriptors and latent topic vectors of tags associated with the images. To select the representative iconic images for the joint clusters, we use a quality ranking learned from a large collection of labeled images. For the purposes of visualization, iconic images are grouped by semantic ldquothemerdquo and multidimensional scaling is used to compute a 2D layout that reflects the relationships between the themes. Results on four large-scale datasets demonstrate the ability of our approach to discover plausible themes and recurring visual motifs for challenging abstract concepts such as ldquoloverdquo and ldquobeautyrdquo.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130805529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
Incremental estimation without specifying a-priori covariance matrices for the novel parameters 增量估计不指定先验协方差矩阵的新参数
C. Beder, Richard Steffen
We will present a novel incremental algorithm for the task of online least-squares estimation. Our approach aims at combining the accuracy of least-squares estimation and the fast computation of recursive estimation techniques like the Kalman filter. Analyzing the structure of least-squares estimation we devise a novel incremental algorithm, which is able to introduce new unknown parameters and observations into an estimation simultaneously and is equivalent to the optimal overall estimation in case of linear models. It constitutes a direct generalization of the well-known Kalman filter allowing to augment the state vector inside the update step. In contrast to classical recursive estimation techniques no artificial initial covariance for the new unknown parameters is required here. We will show, how this new algorithm allows more flexible parameter estimation schemes especially in the case of scene and motion reconstruction from image sequences. Since optimality is not guaranteed in the non-linear case we will also compare our incremental estimation scheme to the optimal bundle adjustment on a real image sequence. It will be shown that competitive results are achievable using the proposed technique.
我们将提出一种新的增量算法用于在线最小二乘估计。我们的方法旨在将最小二乘估计的准确性与卡尔曼滤波等递归估计技术的快速计算相结合。在分析最小二乘估计结构的基础上,提出了一种新的增量估计算法,该算法可以同时将新的未知参数和观测值引入到估计中,相当于线性模型的最优整体估计。它构成了众所周知的卡尔曼滤波的直接推广,允许在更新步骤中增加状态向量。与经典的递归估计技术不同,这里不需要对新的未知参数进行人工初始协方差。我们将展示,这种新算法如何允许更灵活的参数估计方案,特别是在从图像序列中重建场景和运动的情况下。由于在非线性情况下不能保证最优性,我们还将增量估计方案与真实图像序列上的最优束调整进行比较。这将表明,竞争结果是可以实现使用所提出的技术。
{"title":"Incremental estimation without specifying a-priori covariance matrices for the novel parameters","authors":"C. Beder, Richard Steffen","doi":"10.1109/CVPRW.2008.4563139","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563139","url":null,"abstract":"We will present a novel incremental algorithm for the task of online least-squares estimation. Our approach aims at combining the accuracy of least-squares estimation and the fast computation of recursive estimation techniques like the Kalman filter. Analyzing the structure of least-squares estimation we devise a novel incremental algorithm, which is able to introduce new unknown parameters and observations into an estimation simultaneously and is equivalent to the optimal overall estimation in case of linear models. It constitutes a direct generalization of the well-known Kalman filter allowing to augment the state vector inside the update step. In contrast to classical recursive estimation techniques no artificial initial covariance for the new unknown parameters is required here. We will show, how this new algorithm allows more flexible parameter estimation schemes especially in the case of scene and motion reconstruction from image sequences. Since optimality is not guaranteed in the non-linear case we will also compare our incremental estimation scheme to the optimal bundle adjustment on a real image sequence. It will be shown that competitive results are achievable using the proposed technique.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124276834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Multi-parts and multi-feature fusion in face verification 人脸验证中的多部位多特征融合
Yan Xiang, G. Su
Information fusion of multi-biometrics has become a center of focus for biometrics based identification and verification, and there are two fusion categories: intra-modal fusion and multi-modal fusion. In this paper, an intra-modal fusion, that is, multi-parts and multi-feature fusion (MPMFF) for face verification is studied. Two face representations are exploited, including the gray-level intensity feature and Gabor feature. Different from most face recognition methods, the MPMFF method divides a face image into five parts: bare face, eyebrows, eyes, nose and mouth, and different features of the same face part are fused at feature level. Then at decision level, five matching results based on the combined-features of different parts are calculated into a final similar score according to the weighted sum rule. Experiment results on FERET face database and our own face database show that the multi-parts and multi-feature fusion method improves the face verification performance.
多生物特征信息融合已成为基于生物特征识别与验证的研究热点,融合有模态内融合和多模态融合两大类。本文研究了一种用于人脸验证的模内融合,即多部分多特征融合(MPMFF)。利用两种人脸表征,包括灰度强度特征和Gabor特征。与大多数人脸识别方法不同的是,MPMFF方法将人脸图像分为五个部分:裸脸、眉毛、眼睛、鼻子和嘴巴,并在特征层面融合同一面部部位的不同特征。然后在决策层,根据不同部分组合特征的5个匹配结果,根据加权和规则计算出最终的相似分数。在FERET人脸数据库和我们自己的人脸数据库上的实验结果表明,多部分多特征融合方法提高了人脸验证性能。
{"title":"Multi-parts and multi-feature fusion in face verification","authors":"Yan Xiang, G. Su","doi":"10.1109/CVPRW.2008.4563107","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563107","url":null,"abstract":"Information fusion of multi-biometrics has become a center of focus for biometrics based identification and verification, and there are two fusion categories: intra-modal fusion and multi-modal fusion. In this paper, an intra-modal fusion, that is, multi-parts and multi-feature fusion (MPMFF) for face verification is studied. Two face representations are exploited, including the gray-level intensity feature and Gabor feature. Different from most face recognition methods, the MPMFF method divides a face image into five parts: bare face, eyebrows, eyes, nose and mouth, and different features of the same face part are fused at feature level. Then at decision level, five matching results based on the combined-features of different parts are calculated into a final similar score according to the weighted sum rule. Experiment results on FERET face database and our own face database show that the multi-parts and multi-feature fusion method improves the face verification performance.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124447424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Internet video category recognition 互联网视频分类识别
Grant Schindler, Larry Zitnick, Matthew A. Brown
In this paper, we examine the problem of internet video categorization. Specifically, we explore the representation of a video as a ldquobag of wordsrdquo using various combinations of spatial and temporal descriptors. The descriptors incorporate both spatial and temporal gradients as well as optical flow information. We achieve state-of-the-art results on a standard human activity recognition database and demonstrate promising category recognition performance on two new databases of approximately 1000 and 1500 online user-submitted videos, which we will be making available to the community.
本文主要研究网络视频的分类问题。具体来说,我们通过使用空间和时间描述符的各种组合来探索视频作为单词包的表示。描述符包含空间和时间梯度以及光流信息。我们在一个标准的人类活动识别数据库上取得了最先进的结果,并在两个新数据库上展示了有希望的类别识别性能,这些数据库包含大约1000个和1500个在线用户提交的视频,我们将向社区提供这些视频。
{"title":"Internet video category recognition","authors":"Grant Schindler, Larry Zitnick, Matthew A. Brown","doi":"10.1109/CVPRW.2008.4562960","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562960","url":null,"abstract":"In this paper, we examine the problem of internet video categorization. Specifically, we explore the representation of a video as a ldquobag of wordsrdquo using various combinations of spatial and temporal descriptors. The descriptors incorporate both spatial and temporal gradients as well as optical flow information. We achieve state-of-the-art results on a standard human activity recognition database and demonstrate promising category recognition performance on two new databases of approximately 1000 and 1500 online user-submitted videos, which we will be making available to the community.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116952599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Efficient anisotropic α-Kernels decompositions and flows 高效各向异性α-核分解与流动
Micha Feigin-Almon, N. Sochen, B. Vemuri
The Laplacian raised to fractional powers can be used to generate scale spaces as was shown in recent literature. This was later extended for inhomogeneous diffusion processes and more general functions of the Laplacian and studied for the Perona-Malik case. In this paper we extend the results to the truly anisotropic Beltrami flow. We additionally introduce a technique for splitting up the work into smaller patches of the image which greatly reduce the computational complexity and allow for the parallelization of the algorithm. Important issues involved in the numerical implementation are discussed.
拉普拉斯级数的分数次幂可以用来生成尺度空间,这在最近的文献中得到了证明。这后来被推广到非齐次扩散过程和更一般的拉普拉斯函数,并研究了Perona-Malik情况。本文将所得结果推广到真正的各向异性贝尔特拉米流。我们还介绍了一种将工作分割成更小的图像块的技术,这大大降低了计算复杂性,并允许算法的并行化。讨论了数值实现中涉及的重要问题。
{"title":"Efficient anisotropic α-Kernels decompositions and flows","authors":"Micha Feigin-Almon, N. Sochen, B. Vemuri","doi":"10.1109/CVPRW.2008.4562982","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562982","url":null,"abstract":"The Laplacian raised to fractional powers can be used to generate scale spaces as was shown in recent literature. This was later extended for inhomogeneous diffusion processes and more general functions of the Laplacian and studied for the Perona-Malik case. In this paper we extend the results to the truly anisotropic Beltrami flow. We additionally introduce a technique for splitting up the work into smaller patches of the image which greatly reduce the computational complexity and allow for the parallelization of the algorithm. Important issues involved in the numerical implementation are discussed.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123611112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Autonomous navigation and mapping using monocular low-resolution grayscale vision 使用单眼低分辨率灰度视觉的自主导航和制图
Vidya N. Murali, Stan Birchfield
An algorithm is proposed to answer the challenges of autonomous corridor navigation and mapping by a mobile robot equipped with a single forward-facing camera. Using a combination of corridor ceiling lights, visual homing, and entropy, the robot is able to perform straight line navigation down the center of an unknown corridor. Turning at the end of a corridor is accomplished using Jeffrey divergence and time-to-collision, while deflection from dead ends and blank walls uses a scalar entropy measure of the entire image. When combined, these metrics allow the robot to navigate in both textured and untextured environments. The robot can autonomously explore an unknown indoor environment, recovering from difficult situations like corners, blank walls, and initial heading toward a wall. While exploring, the algorithm constructs a Voronoi-based topo-geometric map with nodes representing distinctive places like doors, water fountains, and other corridors. Because the algorithm is based entirely upon low-resolution (32 times 24) grayscale images, processing occurs at over 1000 frames per second.
提出了一种算法来解决配备单个前置摄像头的移动机器人自主走廊导航和地图绘制的挑战。利用走廊顶灯、视觉导航和熵的组合,机器人能够沿着未知走廊的中心进行直线导航。在走廊尽头的转弯使用杰弗里散度和碰撞时间来完成,而从死角和空白墙壁的偏转使用整个图像的标量熵度量。当这些指标结合在一起时,机器人可以在纹理和非纹理环境中导航。该机器人可以自主探索未知的室内环境,从拐角、空白墙壁等困难的情况中恢复过来,并最初朝着墙壁前进。在探索过程中,该算法构建了一个基于voronoi的拓扑几何地图,其中的节点代表不同的地方,如门、喷泉和其他走廊。由于该算法完全基于低分辨率(32 × 24)灰度图像,因此处理速度超过每秒1000帧。
{"title":"Autonomous navigation and mapping using monocular low-resolution grayscale vision","authors":"Vidya N. Murali, Stan Birchfield","doi":"10.1109/CVPRW.2008.4563136","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563136","url":null,"abstract":"An algorithm is proposed to answer the challenges of autonomous corridor navigation and mapping by a mobile robot equipped with a single forward-facing camera. Using a combination of corridor ceiling lights, visual homing, and entropy, the robot is able to perform straight line navigation down the center of an unknown corridor. Turning at the end of a corridor is accomplished using Jeffrey divergence and time-to-collision, while deflection from dead ends and blank walls uses a scalar entropy measure of the entire image. When combined, these metrics allow the robot to navigate in both textured and untextured environments. The robot can autonomously explore an unknown indoor environment, recovering from difficult situations like corners, blank walls, and initial heading toward a wall. While exploring, the algorithm constructs a Voronoi-based topo-geometric map with nodes representing distinctive places like doors, water fountains, and other corridors. Because the algorithm is based entirely upon low-resolution (32 times 24) grayscale images, processing occurs at over 1000 frames per second.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":" 17","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120834117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Integrated segmentation and motion analysis of cardiac MR images using a subject-specific dynamical model 综合分割和运动分析的心脏磁共振图像使用特定主题的动态模型
Yun Zhu, X. Papademetris, A. Sinusas, J. Duncan
In this paper we propose an integrated cardiac segmentation and motion tracking algorithm. First, we present a subject-specific dynamical model (SSDM) that simultaneously handles inter-subject variability and temporal dynamics (intra-subject variability), such that it can progressively identify the subject vector associated with a new cardiac sequence, and use this subject vector to predict the subject-specific segmentation of the future frames based on the shapes observed in earlier frames. Second, we use the segmentation as a guide in selecting feature points with significant shape characteristics, and invoke the generalized robust point matching (G-RPM) strategy with boundary element method (BEM)-based regularization model to estimate physically realistic displacement field in a computationally efficient way. The integrated algorithm is formulated in a recursive Bayesian framework that sequentially segments cardiac images and estimates myocardial displacements. ldquoLeave-one-outrdquo validation on 32 sequences demonstrates that the segmentation results are improved when the SSDM is used, and the tracking results are much more accurate when the segmentation module is added.
本文提出了一种综合的心脏分割和运动跟踪算法。首先,我们提出了一个主体特定的动态模型(SSDM),它可以同时处理主体间的可变性和时间动态(主体内的可变性),这样它就可以逐步识别与新的心脏序列相关的主体向量,并使用这个主体向量来预测基于早期帧中观察到的形状的未来帧的主体特定分割。其次,以分割为指导,选择具有重要形状特征的特征点,并利用基于边界元法(BEM)正则化模型的广义鲁棒点匹配(G-RPM)策略,以高效的计算方式估计物理真实位移场。该集成算法是在一个递归贝叶斯框架中制定的,该框架依次分割心脏图像并估计心肌位移。对32个序列的ldquoleave -one- outdquo验证表明,使用SSDM后,分割结果得到了改善,添加分割模块后,跟踪结果更加准确。
{"title":"Integrated segmentation and motion analysis of cardiac MR images using a subject-specific dynamical model","authors":"Yun Zhu, X. Papademetris, A. Sinusas, J. Duncan","doi":"10.1109/CVPRW.2008.4563007","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563007","url":null,"abstract":"In this paper we propose an integrated cardiac segmentation and motion tracking algorithm. First, we present a subject-specific dynamical model (SSDM) that simultaneously handles inter-subject variability and temporal dynamics (intra-subject variability), such that it can progressively identify the subject vector associated with a new cardiac sequence, and use this subject vector to predict the subject-specific segmentation of the future frames based on the shapes observed in earlier frames. Second, we use the segmentation as a guide in selecting feature points with significant shape characteristics, and invoke the generalized robust point matching (G-RPM) strategy with boundary element method (BEM)-based regularization model to estimate physically realistic displacement field in a computationally efficient way. The integrated algorithm is formulated in a recursive Bayesian framework that sequentially segments cardiac images and estimates myocardial displacements. ldquoLeave-one-outrdquo validation on 32 sequences demonstrates that the segmentation results are improved when the SSDM is used, and the tracking results are much more accurate when the segmentation module is added.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128319264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
IVUS tissue characterization with sub-class error-correcting output codes IVUS组织表征与亚类纠错输出代码
Sergio Escalera, O. Pujol, J. Mauri, P. Radeva
Intravascular ultrasound (IVUS) represents a powerful imaging technique to explore coronary vessels and to study their morphology and histologic properties. In this paper, we characterize different tissues based on Radio Frequency, texture-based, slope-based, and combined features. To deal with the classification of multiple tissues, we require the use of robust multi-class learning techniques. In this context, we propose a strategy to model multi-class classification tasks using sub-classes information in the ECOC framework. The new strategy splits the classes into different subsets according to the applied base classifier. Complex IVUS data sets containing overlapping data are learnt by splitting the original set of classes into sub-classes, and embedding the binary problems in a problem-dependent ECOC design. The method automatically characterizes different tissues, showing performance improvements over the state-of-the-art ECOC techniques for different base classifiers and feature sets.
血管内超声(IVUS)是一种探测冠状血管并研究其形态学和组织学特性的强大成像技术。在本文中,我们基于射频、基于纹理、基于坡度和组合特征来表征不同的组织。为了处理多组织的分类,我们需要使用鲁棒的多类学习技术。在此背景下,我们提出了一种利用ECOC框架中的子类信息对多类分类任务建模的策略。新策略根据应用的基分类器将类划分为不同的子集。包含重叠数据的复杂IVUS数据集通过将原始类集划分为子类来学习,并将二元问题嵌入到问题依赖的ECOC设计中。该方法自动表征不同的组织,在不同的基本分类器和特征集上显示了比最先进的ECOC技术的性能改进。
{"title":"IVUS tissue characterization with sub-class error-correcting output codes","authors":"Sergio Escalera, O. Pujol, J. Mauri, P. Radeva","doi":"10.1109/CVPRW.2008.4563021","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563021","url":null,"abstract":"Intravascular ultrasound (IVUS) represents a powerful imaging technique to explore coronary vessels and to study their morphology and histologic properties. In this paper, we characterize different tissues based on Radio Frequency, texture-based, slope-based, and combined features. To deal with the classification of multiple tissues, we require the use of robust multi-class learning techniques. In this context, we propose a strategy to model multi-class classification tasks using sub-classes information in the ECOC framework. The new strategy splits the classes into different subsets according to the applied base classifier. Complex IVUS data sets containing overlapping data are learnt by splitting the original set of classes into sub-classes, and embedding the binary problems in a problem-dependent ECOC design. The method automatically characterizes different tissues, showing performance improvements over the state-of-the-art ECOC techniques for different base classifiers and feature sets.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128613575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Distributed segmentation and classification of human actions using a wearable motion sensor network 基于可穿戴运动传感器网络的人体动作分布式分割与分类
A. Yang, Sameer Iyengar, S. Sastry, R. Bajcsy, P. Kuryloski, R. Jafari
We propose a distributed recognition method to classify human actions using a low-bandwidth wearable motion sensor network. Given a set of pre-segmented motion sequences as training examples, the algorithm simultaneously segments and classifies human actions, and it also rejects outlying actions that are not in the training set. The classification is distributedly operated on individual sensor nodes and a base station computer. We show that the distribution of multiple action classes satisfies a mixture subspace model, one sub-space for each action class. Given a new test sample, we seek the sparsest linear representation of the sample w.r.t. all training examples. We show that the dominant coefficients in the representation only correspond to the action class of the test sample, and hence its membership is encoded in the representation. We further provide fast linear solvers to compute such representation via l1-minimization. Using up to eight body sensors, the algorithm achieves state-of-the-art 98.8% accuracy on a set of 12 action categories. We further demonstrate that the recognition precision only decreases gracefully using smaller subsets of sensors, which validates the robustness of the distributed framework.
我们提出了一种基于低带宽可穿戴运动传感器网络的分布式识别方法。给定一组预先分割的动作序列作为训练样例,该算法在对人类动作进行分割和分类的同时,也拒绝不在训练集中的外围动作。分类分布在各个传感器节点和基站计算机上。我们证明了多个动作类的分布满足一个混合子空间模型,每个动作类一个子空间。给定一个新的测试样本,我们寻求样本的最稀疏线性表示,w.r.t.所有训练样本。我们表明,表示中的主导系数只对应于测试样本的动作类,因此它的隶属关系被编码在表示中。我们进一步提供了快速的线性求解器,通过11 -最小化来计算这种表示。该算法使用多达8个身体传感器,在12个动作类别中达到了98.8%的准确率。我们进一步证明,使用较小的传感器子集,识别精度只会优雅地降低,这验证了分布式框架的鲁棒性。
{"title":"Distributed segmentation and classification of human actions using a wearable motion sensor network","authors":"A. Yang, Sameer Iyengar, S. Sastry, R. Bajcsy, P. Kuryloski, R. Jafari","doi":"10.1109/CVPRW.2008.4563176","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563176","url":null,"abstract":"We propose a distributed recognition method to classify human actions using a low-bandwidth wearable motion sensor network. Given a set of pre-segmented motion sequences as training examples, the algorithm simultaneously segments and classifies human actions, and it also rejects outlying actions that are not in the training set. The classification is distributedly operated on individual sensor nodes and a base station computer. We show that the distribution of multiple action classes satisfies a mixture subspace model, one sub-space for each action class. Given a new test sample, we seek the sparsest linear representation of the sample w.r.t. all training examples. We show that the dominant coefficients in the representation only correspond to the action class of the test sample, and hence its membership is encoded in the representation. We further provide fast linear solvers to compute such representation via l1-minimization. Using up to eight body sensors, the algorithm achieves state-of-the-art 98.8% accuracy on a set of 12 action categories. We further demonstrate that the recognition precision only decreases gracefully using smaller subsets of sensors, which validates the robustness of the distributed framework.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130036868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 108
Detecting mirror-symmetry of a volumetric shape from its single 2D image 从单个二维图像中检测体积形状的镜像对称性
T. Sawada, Z. Pizlo
We present a new computational model for verifying whether a 3D shape is mirror-symmetric based on its single 2D image. First, a psychophysical experiment which tested human performance in detection of 3D symmetry is described. These psychophysical results led to the formulation of a new algorithm for symmetry detection. The algorithm first recovers the 3D shape using a priori constraints (symmetry, planarity of contours and 3D compactness) and then evaluates the degree of symmetry of the 3D shape. Reliable discrimination by the algorithm between symmetric and asymmetric 3D shapes involves two measures: similarity of the two halves of a 3D shape and compactness of the 3D shape. Performance of this algorithm is highly correlated with that of the subjects. We conclude that this algorithm is a plausible model of the mechanisms used by the human visual system.
我们提出了一个新的计算模型来验证一个三维形状是否基于其单一的二维图像镜像对称。首先,描述了一个心理物理实验,测试了人类在检测三维对称性方面的表现。这些心理物理结果导致了对称检测新算法的形成。该算法首先利用先验约束条件(对称性、轮廓平面性和三维紧凑性)恢复三维形状,然后评估三维形状的对称程度。该算法对对称和非对称三维形状的可靠判别涉及两个指标:三维形状两半的相似度和三维形状的紧凑度。该算法的性能与被试的性能高度相关。我们得出结论,该算法是人类视觉系统使用的机制的合理模型。
{"title":"Detecting mirror-symmetry of a volumetric shape from its single 2D image","authors":"T. Sawada, Z. Pizlo","doi":"10.1109/CVPRW.2008.4562976","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562976","url":null,"abstract":"We present a new computational model for verifying whether a 3D shape is mirror-symmetric based on its single 2D image. First, a psychophysical experiment which tested human performance in detection of 3D symmetry is described. These psychophysical results led to the formulation of a new algorithm for symmetry detection. The algorithm first recovers the 3D shape using a priori constraints (symmetry, planarity of contours and 3D compactness) and then evaluates the degree of symmetry of the 3D shape. Reliable discrimination by the algorithm between symmetric and asymmetric 3D shapes involves two measures: similarity of the two halves of a 3D shape and compactness of the 3D shape. Performance of this algorithm is highly correlated with that of the subjects. We conclude that this algorithm is a plausible model of the mechanisms used by the human visual system.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130173445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1