首页 > 最新文献

IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision最新文献

英文 中文
Coupling video segmentation and action recognition 视频分割与动作识别的耦合
Amir Ghodrati, M. Pedersoli, T. Tuytelaars
Recently a lot of progress has been made in the field of video segmentation. The question then arises whether and how these results can be exploited for this other video processing challenge, action recognition. In this paper we show that a good segmentation is actually very important for recognition. We propose and evaluate several ways to integrate and combine the two tasks: i) recognition using a standard, bottom-up segmentation, ii) using a top-down segmentation geared towards actions, iii) using a segmentation based on inter-video similarities (co-segmentation), and iv) tight integration of recognition and segmentation via iterative learning. Our results clearly show that, on the one hand, the two tasks are interdependent and therefore an iterative optimization of the two makes sense and gives better results. On the other hand, comparable results can also be obtained with two separate steps but mapping the feature-space with a non-linear kernel.
近年来,在视频分割领域取得了很大的进展。接下来的问题是,这些结果是否以及如何被用于另一个视频处理挑战——动作识别。在本文中,我们证明了良好的分割对于识别是非常重要的。我们提出并评估了几种整合和结合这两个任务的方法:i)使用标准的自下而上分割的识别,ii)使用面向动作的自上而下分割,iii)使用基于视频间相似性的分割(共同分割),以及iv)通过迭代学习将识别和分割紧密集成。我们的结果清楚地表明,一方面,这两个任务是相互依赖的,因此对这两个任务进行迭代优化是有意义的,并且会得到更好的结果。另一方面,用非线性核映射特征空间的两个独立步骤也可以得到类似的结果。
{"title":"Coupling video segmentation and action recognition","authors":"Amir Ghodrati, M. Pedersoli, T. Tuytelaars","doi":"10.1109/WACV.2014.6836045","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836045","url":null,"abstract":"Recently a lot of progress has been made in the field of video segmentation. The question then arises whether and how these results can be exploited for this other video processing challenge, action recognition. In this paper we show that a good segmentation is actually very important for recognition. We propose and evaluate several ways to integrate and combine the two tasks: i) recognition using a standard, bottom-up segmentation, ii) using a top-down segmentation geared towards actions, iii) using a segmentation based on inter-video similarities (co-segmentation), and iv) tight integration of recognition and segmentation via iterative learning. Our results clearly show that, on the one hand, the two tasks are interdependent and therefore an iterative optimization of the two makes sense and gives better results. On the other hand, comparable results can also be obtained with two separate steps but mapping the feature-space with a non-linear kernel.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"35 1","pages":"618-625"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72753658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Finger-knuckle-print verification based on vector consistency of corresponding interest points 基于相应兴趣点向量一致性的指关节指纹验证
Min-Ki Kim, P. Flynn
This paper proposes a novel finger-knuckle-print (FKP) verification method based on vector consistency among corresponding interest points (CIPs) detected from aligned finger images. We used two different approaches for reliable detection of CIPs; one method employs SIFT features and captures gradient directionality, and the other method employs phase correlation to represent the intensity field surrounding an interest point. The consistency of interframe displacements between pairs of matching CIPs in a match pair is used as a matching score. Such displacements will show consistency in a genuine match but not in an impostor match. Experimental results show that the proposed approach is effective in FKP verification.
本文提出了一种基于从对齐的手指图像中检测到的相应兴趣点(cip)之间的向量一致性的指关节指纹(FKP)验证方法。我们使用了两种不同的方法来可靠地检测cip;一种方法利用SIFT特征捕获梯度方向性,另一种方法利用相位相关表示兴趣点周围的强度场。匹配对中匹配cip对之间帧间位移的一致性被用作匹配分数。这样的位移将在真正的匹配中显示一致性,但在冒牌货匹配中则不然。实验结果表明,该方法在FKP验证中是有效的。
{"title":"Finger-knuckle-print verification based on vector consistency of corresponding interest points","authors":"Min-Ki Kim, P. Flynn","doi":"10.1109/WACV.2014.6835996","DOIUrl":"https://doi.org/10.1109/WACV.2014.6835996","url":null,"abstract":"This paper proposes a novel finger-knuckle-print (FKP) verification method based on vector consistency among corresponding interest points (CIPs) detected from aligned finger images. We used two different approaches for reliable detection of CIPs; one method employs SIFT features and captures gradient directionality, and the other method employs phase correlation to represent the intensity field surrounding an interest point. The consistency of interframe displacements between pairs of matching CIPs in a match pair is used as a matching score. Such displacements will show consistency in a genuine match but not in an impostor match. Experimental results show that the proposed approach is effective in FKP verification.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"15 1","pages":"992-997"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76430927","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Linear Local Distance coding for classification of HEp-2 staining patterns 线性局部距离编码用于HEp-2染色模式分类
Xiang Xu, F. Lin, Carol Ng, K. Leong
Indirect Immunofluorescence (IIF) on Human Epithelial-2 (HEp-2) cells is the recommended methodology for detecting some specific autoimmune diseases by searching for antinuclear antibodies (ANAs) within a patient's serum. Due to the limitations of IIF such as subjective evaluation, automated Computer-Aided Diagnosis (CAD) system is required for diagnostic purposes. In particular, staining patterns classification of HEp-2 cells is a challenging task. In this paper, we adopt a feature extraction-coding-pooling framework which has shown impressive performance in image classification tasks, because it can obtain discriminative and effective image representation. However, the information loss is inevitable in the coding process. Therefore, we propose a Linear Local Distance (LLD) coding method to capture more discriminative information. LLD transforms original local feature to local distance vector by searching for local nearest few neighbors of local feature in the class-specific manifolds. The obtained local distance vector is further encoded and pooled together to get salient image representation. We demonstrate the effectiveness of LLD method on a public HEp-2 cells dataset containing six major staining patterns. Experimental results show that our approach has a superior performance to the state-of-the-art coding methods for staining patterns classification of HEp-2 cells.
人上皮-2 (HEp-2)细胞的间接免疫荧光(IIF)是通过在患者血清中寻找抗核抗体(ANAs)来检测某些特定自身免疫性疾病的推荐方法。由于IIF的局限性,如主观评价,需要自动计算机辅助诊断(CAD)系统进行诊断。特别是HEp-2细胞的染色模式分类是一项具有挑战性的任务。在本文中,我们采用了一种特征提取-编码池框架,该框架可以获得判别和有效的图像表示,在图像分类任务中表现出令人印象深刻的性能。然而,在编码过程中,信息丢失是不可避免的。因此,我们提出了一种线性局部距离(LLD)编码方法来捕获更多的判别信息。LLD通过在特定类流形中搜索局部特征的局部最近近邻,将原始局部特征转化为局部距离向量。对得到的局部距离向量进行进一步编码和汇总,得到显著图像表示。我们证明了LLD方法在包含六种主要染色模式的公共HEp-2细胞数据集上的有效性。实验结果表明,我们的方法在HEp-2细胞染色模式分类方面具有优于目前最先进的编码方法的性能。
{"title":"Linear Local Distance coding for classification of HEp-2 staining patterns","authors":"Xiang Xu, F. Lin, Carol Ng, K. Leong","doi":"10.1109/WACV.2014.6836073","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836073","url":null,"abstract":"Indirect Immunofluorescence (IIF) on Human Epithelial-2 (HEp-2) cells is the recommended methodology for detecting some specific autoimmune diseases by searching for antinuclear antibodies (ANAs) within a patient's serum. Due to the limitations of IIF such as subjective evaluation, automated Computer-Aided Diagnosis (CAD) system is required for diagnostic purposes. In particular, staining patterns classification of HEp-2 cells is a challenging task. In this paper, we adopt a feature extraction-coding-pooling framework which has shown impressive performance in image classification tasks, because it can obtain discriminative and effective image representation. However, the information loss is inevitable in the coding process. Therefore, we propose a Linear Local Distance (LLD) coding method to capture more discriminative information. LLD transforms original local feature to local distance vector by searching for local nearest few neighbors of local feature in the class-specific manifolds. The obtained local distance vector is further encoded and pooled together to get salient image representation. We demonstrate the effectiveness of LLD method on a public HEp-2 cells dataset containing six major staining patterns. Experimental results show that our approach has a superior performance to the state-of-the-art coding methods for staining patterns classification of HEp-2 cells.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"8 1","pages":"393-400"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81445168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Structure-aware keypoint tracking for partial occlusion handling 部分遮挡处理的结构感知关键点跟踪
W. Bouachir, Guillaume-Alexandre Bilodeau
This paper introduces a novel keypoint-based method for visual object tracking. To represent the target, we use a new model combining color distribution with keypoints. The appearance model also incorporates the spatial layout of the keypoints, encoding the object structure learned during tracking. With this multi-feature appearance model, our Structure-Aware Tracker (SAT) estimates accurately the target location using three main steps. First, the search space is reduced to the most likely image regions with a probabilistic approach. Second, the target location is estimated in the reduced search space using deterministic keypoint matching. Finally, the location prediction is corrected by exploiting the keypoint structural model with a voting-based method. By applying our SAT on several tracking problems, we show that location correction based on structural constraints is a key technique to improve prediction in moderately crowded scenes, even if only a small part of the target is visible. We also conduct comparison with a number of state-of-the-art trackers and demonstrate the competitiveness of the proposed method.
提出了一种新的基于关键点的视觉目标跟踪方法。为了表示目标,我们使用了一种结合颜色分布和关键点的新模型。外观模型还包含关键点的空间布局,对跟踪过程中学习到的对象结构进行编码。利用这种多特征外观模型,我们的结构感知跟踪器(SAT)通过三个主要步骤准确地估计目标位置。首先,用概率方法将搜索空间缩小到最可能的图像区域;其次,利用确定性关键点匹配在简化后的搜索空间中估计目标位置。最后,采用基于投票的方法利用关键点结构模型对位置预测进行修正。通过将我们的SAT应用于几个跟踪问题,我们表明,在中等拥挤的场景中,即使只有一小部分目标可见,基于结构约束的位置校正是提高预测能力的关键技术。我们还与一些最先进的跟踪器进行了比较,并证明了所提出方法的竞争力。
{"title":"Structure-aware keypoint tracking for partial occlusion handling","authors":"W. Bouachir, Guillaume-Alexandre Bilodeau","doi":"10.1109/WACV.2014.6836011","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836011","url":null,"abstract":"This paper introduces a novel keypoint-based method for visual object tracking. To represent the target, we use a new model combining color distribution with keypoints. The appearance model also incorporates the spatial layout of the keypoints, encoding the object structure learned during tracking. With this multi-feature appearance model, our Structure-Aware Tracker (SAT) estimates accurately the target location using three main steps. First, the search space is reduced to the most likely image regions with a probabilistic approach. Second, the target location is estimated in the reduced search space using deterministic keypoint matching. Finally, the location prediction is corrected by exploiting the keypoint structural model with a voting-based method. By applying our SAT on several tracking problems, we show that location correction based on structural constraints is a key technique to improve prediction in moderately crowded scenes, even if only a small part of the target is visible. We also conduct comparison with a number of state-of-the-art trackers and demonstrate the competitiveness of the proposed method.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"28 1","pages":"877-884"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90531344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Multi-view action recognition one camera at a time 多视角动作识别,一次一个摄像头
Scott Spurlock, Richard Souvenir
For human action recognition methods, there is often a trade-off between classification accuracy and computational efficiency. Methods that include 3D information from multiple cameras are often computationally expensive and not suitable for real-time application. 2D, frame-based methods are generally more efficient, but suffer from lower recognition accuracies. In this paper, we present a hybrid keypose-based method that operates in a multi-camera environment, but uses only a single camera at a time. We learn, for each keypose, the relative utility of a particular viewpoint compared with switching to a different available camera in the network for future classification. On a benchmark multi-camera action recognition dataset, our method outperforms approaches that incorporate all available cameras.
对于人体动作识别方法,通常需要在分类精度和计算效率之间进行权衡。包含来自多个摄像机的3D信息的方法通常在计算上很昂贵,并且不适合实时应用。基于帧的二维方法通常效率更高,但识别精度较低。在本文中,我们提出了一种基于密钥的混合方法,该方法可以在多相机环境中运行,但每次只使用一个相机。我们学习,对于每个键位,一个特定视点的相对效用,与切换到网络中不同的可用摄像机进行未来分类相比。在一个基准的多相机动作识别数据集上,我们的方法优于包含所有可用相机的方法。
{"title":"Multi-view action recognition one camera at a time","authors":"Scott Spurlock, Richard Souvenir","doi":"10.1109/WACV.2014.6836047","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836047","url":null,"abstract":"For human action recognition methods, there is often a trade-off between classification accuracy and computational efficiency. Methods that include 3D information from multiple cameras are often computationally expensive and not suitable for real-time application. 2D, frame-based methods are generally more efficient, but suffer from lower recognition accuracies. In this paper, we present a hybrid keypose-based method that operates in a multi-camera environment, but uses only a single camera at a time. We learn, for each keypose, the relative utility of a particular viewpoint compared with switching to a different available camera in the network for future classification. On a benchmark multi-camera action recognition dataset, our method outperforms approaches that incorporate all available cameras.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"91 1","pages":"604-609"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79517106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Interactively test driving an object detector: Estimating performance on unlabeled data 交互式测试驱动对象检测器:评估未标记数据上的性能
Rushil Anirudh, P. Turaga
In this paper, we study the problem of `test-driving' a detector, i.e. allowing a human user to get a quick sense of how well the detector generalizes to their specific requirement. To this end, we present the first system that estimates detector performance interactively without extensive ground truthing using a human in the loop. We approach this as a problem of estimating proportions and show that it is possible to make accurate inferences on the proportion of classes or groups within a large data collection by observing only 5 - 10% of samples from the data. In estimating the false detections (for precision), the samples are chosen carefully such that the overall characteristics of the data collection are preserved. Next, inspired by its use in estimating disease propagation we apply pooled testing approaches to estimate missed detections (for recall) from the dataset. The estimates thus obtained are close to the ones obtained using ground truth, thus reducing the need for extensive labeling which is expensive and time consuming.
在本文中,我们研究了“试驾”检测器的问题,即允许人类用户快速了解检测器对其特定需求的泛化程度。为此,我们提出了第一个系统,该系统可以交互式地估计探测器的性能,而不需要在环路中使用人类进行广泛的地面真实性。我们将此视为估计比例的问题,并表明仅通过观察数据中5 - 10%的样本,就可以准确推断出大型数据集合中类别或组的比例。在估计误检(为了精度)时,要仔细选择样本,以便保留数据收集的总体特征。接下来,受其在估计疾病传播中的应用的启发,我们应用池测试方法来估计数据集中的未检出(召回率)。由此获得的估计值与使用地面真值获得的估计值接近,从而减少了昂贵且耗时的大量标记的需要。
{"title":"Interactively test driving an object detector: Estimating performance on unlabeled data","authors":"Rushil Anirudh, P. Turaga","doi":"10.1109/WACV.2014.6836104","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836104","url":null,"abstract":"In this paper, we study the problem of `test-driving' a detector, i.e. allowing a human user to get a quick sense of how well the detector generalizes to their specific requirement. To this end, we present the first system that estimates detector performance interactively without extensive ground truthing using a human in the loop. We approach this as a problem of estimating proportions and show that it is possible to make accurate inferences on the proportion of classes or groups within a large data collection by observing only 5 - 10% of samples from the data. In estimating the false detections (for precision), the samples are chosen carefully such that the overall characteristics of the data collection are preserved. Next, inspired by its use in estimating disease propagation we apply pooled testing approaches to estimate missed detections (for recall) from the dataset. The estimates thus obtained are close to the ones obtained using ground truth, thus reducing the need for extensive labeling which is expensive and time consuming.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"91 1","pages":"175-182"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87849863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Comparison of face detection and image classification for detecting front seat passengers in vehicles 车辆前座乘客人脸检测与图像分类的比较
Y. Artan, P. Paul, F. Perronnin, A. Burry
Due to the high volume of traffic on modern roadways, transportation agencies have proposed High Occupancy Vehicle (HOV) lanes and High Occupancy Tolling (HOT) lanes to promote car pooling. However, enforcement of the rules of these lanes is currently performed by roadside enforcement officers using visual observation. Manual roadside enforcement is known to be inefficient, costly, potentially dangerous, and ultimately ineffective. Violation rates up to 50%-80% have been reported, while manual enforcement rates of less than 10% are typical. Therefore, there is a need for automated vehicle occupancy detection to support HOV/HOT lane enforcement. A key component of determining vehicle occupancy is to determine whether or not the vehicle's front passenger seat is occupied. In this paper, we examine two methods of determining vehicle front seat occupancy using a near infrared (NIR) camera system pointed at the vehicle's front windshield. The first method examines a state-of-the-art deformable part model (DPM) based face detection system that is robust to facial pose. The second method examines state-of-the-art local aggregation based image classification using bag-of-visual-words (BOW) and Fisher vectors (FV). A dataset of 3000 images was collected on a public roadway and is used to perform the comparison. From these experiments it is clear that the image classification approach is superior for this problem.
由于现代道路上的交通量很大,交通运输机构提出了高占用车辆(HOV)车道和高占用收费(HOT)车道来促进拼车。然而,这些车道规则的执行目前是由路边执法人员通过目视观察来执行的。众所周知,人工路边执法效率低、成本高、有潜在危险,而且最终无效。据报道,违规率高达50%-80%,而人工执法率通常低于10%。因此,需要自动车辆占用检测来支持HOV/HOT车道执法。确定车辆占用率的一个关键组成部分是确定车辆的前排乘客座位是否被占用。在本文中,我们研究了两种确定车辆前座占用率的方法,使用近红外(NIR)相机系统指向车辆的前挡风玻璃。第一种方法研究了一种基于最先进的可变形部件模型(DPM)的面部检测系统,该系统对面部姿势具有鲁棒性。第二种方法使用视觉词袋(BOW)和费雪向量(FV)检查最先进的基于局部聚合的图像分类。在公共道路上收集了3000张图像的数据集,并用于进行比较。从这些实验中可以清楚地看出,图像分类方法在这个问题上是优越的。
{"title":"Comparison of face detection and image classification for detecting front seat passengers in vehicles","authors":"Y. Artan, P. Paul, F. Perronnin, A. Burry","doi":"10.1109/WACV.2014.6835994","DOIUrl":"https://doi.org/10.1109/WACV.2014.6835994","url":null,"abstract":"Due to the high volume of traffic on modern roadways, transportation agencies have proposed High Occupancy Vehicle (HOV) lanes and High Occupancy Tolling (HOT) lanes to promote car pooling. However, enforcement of the rules of these lanes is currently performed by roadside enforcement officers using visual observation. Manual roadside enforcement is known to be inefficient, costly, potentially dangerous, and ultimately ineffective. Violation rates up to 50%-80% have been reported, while manual enforcement rates of less than 10% are typical. Therefore, there is a need for automated vehicle occupancy detection to support HOV/HOT lane enforcement. A key component of determining vehicle occupancy is to determine whether or not the vehicle's front passenger seat is occupied. In this paper, we examine two methods of determining vehicle front seat occupancy using a near infrared (NIR) camera system pointed at the vehicle's front windshield. The first method examines a state-of-the-art deformable part model (DPM) based face detection system that is robust to facial pose. The second method examines state-of-the-art local aggregation based image classification using bag-of-visual-words (BOW) and Fisher vectors (FV). A dataset of 3000 images was collected on a public roadway and is used to perform the comparison. From these experiments it is clear that the image classification approach is superior for this problem.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"1 1","pages":"1006-1012"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74341594","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Scale-Space SIFT flow 尺度空间SIFT流
Weichao Qiu, Xinggang Wang, X. Bai, A. Yuille, Z. Tu
The state-of-the-art SIFT flow has been widely adopted for the general image matching task, especially in dealing with image pairs from similar scenes but with different object configurations. However, the way in which the dense SIFT features are computed at a fixed scale in the SIFT flow method limits its capability of dealing with scenes of large scale changes. In this paper, we propose a simple, intuitive, and very effective approach, Scale-Space SIFT flow, to deal with the large scale differences in different image locations. We introduce a scale field to the SIFT flow function to automatically explore the scale deformations. Our approach achieves similar performance as the SIFT flow method on general natural scenes but obtains significant improvement on the images with large scale differences. Compared with a recent method that addresses the similar problem, our approach shows its clear advantage being more effective, and significantly less demanding in memory and time requirement.
最先进的SIFT流已被广泛应用于一般的图像匹配任务,特别是在处理来自相似场景但具有不同目标配置的图像对时。然而,SIFT流方法中密集SIFT特征在固定尺度上的计算方式限制了其处理大规模变化场景的能力。在本文中,我们提出了一种简单、直观且非常有效的方法——尺度-空间SIFT流来处理不同图像位置的大尺度差异。我们在SIFT流函数中引入尺度场来自动探测尺度变形。我们的方法在一般的自然场景上取得了与SIFT流方法相似的性能,但在大尺度差异的图像上取得了显著的改进。与最近解决类似问题的方法相比,我们的方法显示出其明显的优势,即更有效,并且对内存和时间的要求显着降低。
{"title":"Scale-Space SIFT flow","authors":"Weichao Qiu, Xinggang Wang, X. Bai, A. Yuille, Z. Tu","doi":"10.1109/WACV.2014.6835734","DOIUrl":"https://doi.org/10.1109/WACV.2014.6835734","url":null,"abstract":"The state-of-the-art SIFT flow has been widely adopted for the general image matching task, especially in dealing with image pairs from similar scenes but with different object configurations. However, the way in which the dense SIFT features are computed at a fixed scale in the SIFT flow method limits its capability of dealing with scenes of large scale changes. In this paper, we propose a simple, intuitive, and very effective approach, Scale-Space SIFT flow, to deal with the large scale differences in different image locations. We introduce a scale field to the SIFT flow function to automatically explore the scale deformations. Our approach achieves similar performance as the SIFT flow method on general natural scenes but obtains significant improvement on the images with large scale differences. Compared with a recent method that addresses the similar problem, our approach shows its clear advantage being more effective, and significantly less demanding in memory and time requirement.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"4 1","pages":"1112-1119"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75073359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Extending explicit shape regression with mixed feature channels and pose priors 使用混合特征通道和位姿先验扩展显式形状回归
Matthias Richter, Hua Gao, H. K. Ekenel
Facial feature detection offers a wide range of applications, e.g. in facial image processing, human computer interaction, consumer electronics, and the entertainment industry. These applications impose two antagonistic key requirements: high processing speed and high detection accuracy. We address both by expanding upon the recently proposed explicit shape regression [1] to (a) allow usage and mixture of different feature channels, and (b) include head pose information to improve detection performance in non-cooperative environments. Using the publicly available “wild” datasets LFW [10] and AFLW [11], we show that using these extensions outperforms the baseline (up to 10% gain in accuracy at 8% IOD) as well as other state-of-the-art methods.
面部特征检测提供了广泛的应用,例如面部图像处理、人机交互、消费电子和娱乐行业。这些应用有两个关键的要求:高处理速度和高检测精度。我们通过扩展最近提出的显式形状回归[1]来解决这两个问题,以(a)允许使用和混合不同的特征通道,以及(b)包括头部姿势信息以提高非合作环境中的检测性能。使用公开可用的“野生”数据集LFW[10]和AFLW[11],我们表明使用这些扩展优于基线(在8% IOD下精度提高10%)以及其他最先进的方法。
{"title":"Extending explicit shape regression with mixed feature channels and pose priors","authors":"Matthias Richter, Hua Gao, H. K. Ekenel","doi":"10.1109/WACV.2014.6835993","DOIUrl":"https://doi.org/10.1109/WACV.2014.6835993","url":null,"abstract":"Facial feature detection offers a wide range of applications, e.g. in facial image processing, human computer interaction, consumer electronics, and the entertainment industry. These applications impose two antagonistic key requirements: high processing speed and high detection accuracy. We address both by expanding upon the recently proposed explicit shape regression [1] to (a) allow usage and mixture of different feature channels, and (b) include head pose information to improve detection performance in non-cooperative environments. Using the publicly available “wild” datasets LFW [10] and AFLW [11], we show that using these extensions outperforms the baseline (up to 10% gain in accuracy at 8% IOD) as well as other state-of-the-art methods.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"71 7","pages":"1013-1019"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/WACV.2014.6835993","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72454340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Discovering discriminative cell attributes for HEp-2 specimen image classification 发现HEp-2标本图像分类的鉴别细胞属性
A. Wiliem, Peter Hobson, B. Lovell
Recently, there has been a growing interest in developing Computer Aided Diagnostic (CAD) systems for improving the reliability and consistency of pathology test results. This paper describes a novel CAD system for the Anti-Nuclear Antibody (ANA) test via Indirect Immunofluorescence protocol on Human Epithelial Type 2 (HEp-2) cells. While prior works have primarily focused on classifying cell images extracted from ANA specimen images, this work takes a further step by focussing on the specimen image classification problem itself. Our system is able to efficiently classify specimen images as well as producing meaningful descriptions of ANA pattern class which helps physicians to understand the differences between various ANA patterns. We achieve this goal by designing a specimen-level image descriptor that: (1) is highly discriminative; (2) has small descriptor length and (3) is semantically meaningful at the cell level. In our work, a specimen image descriptor is represented by its overall cell attribute descriptors. As such, we propose two max-margin based learning schemes to discover cell attributes whilst still maintaining the discrimination of the specimen image descriptor. Our learning schemes differ from the existing discriminative attribute learning approaches as they primarily focus on discovering image-level attributes. Comparative evaluations were undertaken to contrast the proposed approach to various state-of-the-art approaches on a novel HEp-2 cell dataset which was specifically proposed for the specimen-level classification. Finally, we showcase the ability of the proposed approach to provide textual descriptions to explain ANA patterns.
最近,人们对开发计算机辅助诊断(CAD)系统以提高病理检测结果的可靠性和一致性越来越感兴趣。本文介绍了一种新的CAD系统,用于间接免疫荧光法检测人上皮细胞2型(HEp-2)的抗核抗体(ANA)。虽然之前的工作主要集中在对从ANA样本图像中提取的细胞图像进行分类,但这项工作进一步关注了样本图像分类问题本身。我们的系统能够有效地对标本图像进行分类,并产生有意义的ANA模式类描述,这有助于医生理解各种ANA模式之间的差异。我们通过设计一个样本级的图像描述符来实现这一目标,该描述符:(1)具有高判别性;(2)描述符长度较小,(3)在单元级具有语义意义。在我们的工作中,标本图像描述符由其整体细胞属性描述符表示。因此,我们提出了两种基于最大边界的学习方案来发现细胞属性,同时仍然保持样本图像描述符的区分。我们的学习方案不同于现有的判别属性学习方法,因为它们主要侧重于发现图像级属性。进行了比较评估,将所提出的方法与专门为标本水平分类提出的新型HEp-2细胞数据集上的各种最新方法进行对比。最后,我们展示了所建议的方法提供文本描述来解释ANA模式的能力。
{"title":"Discovering discriminative cell attributes for HEp-2 specimen image classification","authors":"A. Wiliem, Peter Hobson, B. Lovell","doi":"10.1109/WACV.2014.6836071","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836071","url":null,"abstract":"Recently, there has been a growing interest in developing Computer Aided Diagnostic (CAD) systems for improving the reliability and consistency of pathology test results. This paper describes a novel CAD system for the Anti-Nuclear Antibody (ANA) test via Indirect Immunofluorescence protocol on Human Epithelial Type 2 (HEp-2) cells. While prior works have primarily focused on classifying cell images extracted from ANA specimen images, this work takes a further step by focussing on the specimen image classification problem itself. Our system is able to efficiently classify specimen images as well as producing meaningful descriptions of ANA pattern class which helps physicians to understand the differences between various ANA patterns. We achieve this goal by designing a specimen-level image descriptor that: (1) is highly discriminative; (2) has small descriptor length and (3) is semantically meaningful at the cell level. In our work, a specimen image descriptor is represented by its overall cell attribute descriptors. As such, we propose two max-margin based learning schemes to discover cell attributes whilst still maintaining the discrimination of the specimen image descriptor. Our learning schemes differ from the existing discriminative attribute learning approaches as they primarily focus on discovering image-level attributes. Comparative evaluations were undertaken to contrast the proposed approach to various state-of-the-art approaches on a novel HEp-2 cell dataset which was specifically proposed for the specimen-level classification. Finally, we showcase the ability of the proposed approach to provide textual descriptions to explain ANA patterns.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"11 1","pages":"423-430"},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78498664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
期刊
IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1