首页 > 最新文献

2014 22nd International Conference on Pattern Recognition最新文献

英文 中文
Learning Convolutional Nonlinear Features for K Nearest Neighbor Image Classification 卷积非线性特征在K近邻图像分类中的学习
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.746
Weiqiang Ren, Yinan Yu, Junge Zhang, Kaiqi Huang
Learning low-dimensional feature representations is a crucial task in machine learning and computer vision. Recently the impressive breakthrough in general object recognition made by large scale convolutional networks shows that convolutional networks are able to extract discriminative hierarchical features in large scale object classification task. However, for vision tasks other than end-to-end classification, such as K Nearest Neighbor classification, the learned intermediate features are not necessary optimal for the specific problem. In this paper, we aim to exploit the power of deep convolutional networks and optimize the output feature layer with respect to the task of K Nearest Neighbor (kNN) classification. By directly optimizing the kNN classification error on training data, we in fact learn convolutional nonlinear features in a data-driven and task-driven way. Experimental results on standard image classification benchmarks show that the proposed method is able to learn better feature representations than other general end-to-end classification methods on kNN classification task.
学习低维特征表示是机器学习和计算机视觉中的一项重要任务。近年来,大规模卷积网络在一般目标识别方面取得了令人印象深刻的突破,这表明卷积网络能够在大规模目标分类任务中提取有区别的层次特征。然而,对于端到端分类以外的视觉任务,例如K近邻分类,学习到的中间特征对于特定问题来说不一定是最优的。在本文中,我们的目标是利用深度卷积网络的强大功能,并针对K最近邻(kNN)分类任务优化输出特征层。通过直接优化训练数据上的kNN分类误差,我们实际上是以数据驱动和任务驱动的方式学习卷积非线性特征。在标准图像分类基准上的实验结果表明,在kNN分类任务上,该方法比其他一般的端到端分类方法能够更好地学习到特征表示。
{"title":"Learning Convolutional Nonlinear Features for K Nearest Neighbor Image Classification","authors":"Weiqiang Ren, Yinan Yu, Junge Zhang, Kaiqi Huang","doi":"10.1109/ICPR.2014.746","DOIUrl":"https://doi.org/10.1109/ICPR.2014.746","url":null,"abstract":"Learning low-dimensional feature representations is a crucial task in machine learning and computer vision. Recently the impressive breakthrough in general object recognition made by large scale convolutional networks shows that convolutional networks are able to extract discriminative hierarchical features in large scale object classification task. However, for vision tasks other than end-to-end classification, such as K Nearest Neighbor classification, the learned intermediate features are not necessary optimal for the specific problem. In this paper, we aim to exploit the power of deep convolutional networks and optimize the output feature layer with respect to the task of K Nearest Neighbor (kNN) classification. By directly optimizing the kNN classification error on training data, we in fact learn convolutional nonlinear features in a data-driven and task-driven way. Experimental results on standard image classification benchmarks show that the proposed method is able to learn better feature representations than other general end-to-end classification methods on kNN classification task.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"178 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123571961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Real-Time Tracking via Deformable Structure Regression Learning 基于可变形结构回归学习的实时跟踪
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.379
Xian Yang, Quan Xiao, Shoujue Wang, Peizhong Liu
Visual object tracking is a challenging task because designing an effective and efficient appearance model is difficult. Current online tracking algorithms treat tracking as a classification task and use labeled samples to update appearance model. However, it is not clear to evaluate instance confidence belong to the object. In this paper, we propose a simple and efficient tracking algorithm with a deformable structure appearance. In our method, model updates with continuous labeled samples which are dense sampling. In order to improve the accuracy, we introduce a couple-layer regression model which prevents negative background from impacting on the model learning rather than traditional classification. The proposed DSR tracker runs in real-time and performs favorably against state-of-the-art trackers on various challenging sequences.
视觉目标跟踪是一项具有挑战性的任务,因为设计一个有效和高效的外观模型是困难的。目前的在线跟踪算法将跟踪作为一种分类任务,使用标记样本更新外观模型。但是,不清楚如何评估属于对象的实例置信度。在本文中,我们提出了一种简单有效的具有可变形结构外观的跟踪算法。在我们的方法中,使用连续的标记样本进行模型更新,这是密集采样。为了提高准确率,我们引入了一种防止负面背景影响模型学习的双层回归模型,而不是传统的分类方法。所提出的DSR跟踪器可以实时运行,并且在各种具有挑战性的序列上优于最先进的跟踪器。
{"title":"Real-Time Tracking via Deformable Structure Regression Learning","authors":"Xian Yang, Quan Xiao, Shoujue Wang, Peizhong Liu","doi":"10.1109/ICPR.2014.379","DOIUrl":"https://doi.org/10.1109/ICPR.2014.379","url":null,"abstract":"Visual object tracking is a challenging task because designing an effective and efficient appearance model is difficult. Current online tracking algorithms treat tracking as a classification task and use labeled samples to update appearance model. However, it is not clear to evaluate instance confidence belong to the object. In this paper, we propose a simple and efficient tracking algorithm with a deformable structure appearance. In our method, model updates with continuous labeled samples which are dense sampling. In order to improve the accuracy, we introduce a couple-layer regression model which prevents negative background from impacting on the model learning rather than traditional classification. The proposed DSR tracker runs in real-time and performs favorably against state-of-the-art trackers on various challenging sequences.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"161 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114054945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Optimal Performance-Efficiency Trade-off for Bag of Words Classification of Road Signs 道路标志词袋分类的最优性能-效率权衡
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.517
L. Hazelhoff, Ivo M. Creusen, P. D. With
This paper focuses on road sign classification for creating accurate and up-to-date inventories of traffic signs, which is important for road safety and maintenance. This is a challenging multi-class classification task, as a large number of different sign types exist which only differ in minor details. Moreover, changes in viewpoint, capturing conditions and partial occlusions result in large intra-class variations. Ideally, road sign classification systems should be robust against these variations, while having an acceptable computational load. This paper presents a classification approach based on the popular Bag Of Words (BOW) framework, which we optimize towards the best trade-off between performance and execution time. We analyze the performance aspects of PCA-based dimensionality reduction, soft and hard assignment for BOW codebook matching and the codebook size. Furthermore, we provide an efficient implementation scheme. We compare these techniques to design a fast and accurate BOW-based classification scheme. This approach allows for the selection of a fast but accurate classification methodology. This BOW approach is compared against structural classification, and we show that their combination outperforms both individual methods. This combination, exploiting both BOW and structural information, attains high classification scores (96.25% to 98%) on our challenging real-world datasets.
本文的重点是道路标志分类,以创建准确和最新的交通标志清单,这对道路安全和维护是重要的。这是一个具有挑战性的多类别分类任务,因为存在大量不同的标志类型,只是在小细节上有所不同。此外,视点、捕获条件和部分遮挡的变化会导致很大的类内变化。理想情况下,道路标志分类系统应该对这些变化具有鲁棒性,同时具有可接受的计算负载。本文提出了一种基于流行的词包(BOW)框架的分类方法,我们对其进行了优化,以实现性能和执行时间之间的最佳权衡。分析了基于pca的降维算法、BOW码本匹配的软、硬赋值算法和码本大小算法的性能。此外,我们还提供了一个有效的实现方案。我们比较了这些技术,设计了一个快速准确的基于bow的分类方案。这种方法允许选择快速但准确的分类方法。将这种BOW方法与结构分类方法进行了比较,结果表明它们的组合优于两种单独的方法。这种结合利用了BOW和结构信息,在具有挑战性的真实数据集上获得了很高的分类分数(96.25%到98%)。
{"title":"Optimal Performance-Efficiency Trade-off for Bag of Words Classification of Road Signs","authors":"L. Hazelhoff, Ivo M. Creusen, P. D. With","doi":"10.1109/ICPR.2014.517","DOIUrl":"https://doi.org/10.1109/ICPR.2014.517","url":null,"abstract":"This paper focuses on road sign classification for creating accurate and up-to-date inventories of traffic signs, which is important for road safety and maintenance. This is a challenging multi-class classification task, as a large number of different sign types exist which only differ in minor details. Moreover, changes in viewpoint, capturing conditions and partial occlusions result in large intra-class variations. Ideally, road sign classification systems should be robust against these variations, while having an acceptable computational load. This paper presents a classification approach based on the popular Bag Of Words (BOW) framework, which we optimize towards the best trade-off between performance and execution time. We analyze the performance aspects of PCA-based dimensionality reduction, soft and hard assignment for BOW codebook matching and the codebook size. Furthermore, we provide an efficient implementation scheme. We compare these techniques to design a fast and accurate BOW-based classification scheme. This approach allows for the selection of a fast but accurate classification methodology. This BOW approach is compared against structural classification, and we show that their combination outperforms both individual methods. This combination, exploiting both BOW and structural information, attains high classification scores (96.25% to 98%) on our challenging real-world datasets.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114744048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Interactive Design of Object Classifiers in Remote Sensing 遥感目标分类器的交互设计
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.444
B. L. Saux
This paper deals with the interactive design of generic classifiers for aerial images. In many real-life cases, object detectors that work are not available, due to a new geographical context or a need for a type of object unseen before. We propose an approach for on-line learning of such detectors using user interactions. Variants of gradient boosting and support-vector machine classification are proposed to cope with the problems raised by interactivity: unbalanced and partially mislabeled training data. We assess our framework for various visual classes (buildings, vegetation, cars, visual changes) on challenging data corresponding to several applications (SAR or optical sensors at various resolutions). We show that our model and algorithms outperform several state-of-the-art baselines for feature extraction and learning in remote sensing.
本文研究了航空图像通用分类器的交互设计。在许多现实生活中,由于新的地理环境或对以前未见过的物体类型的需求,无法使用可用的物体探测器。我们提出了一种使用用户交互在线学习这种检测器的方法。提出了梯度增强和支持向量机分类的变体,以应对交互性带来的问题:不平衡和部分错误标记的训练数据。我们对不同视觉类别(建筑、植被、汽车、视觉变化)的框架进行评估,并对不同应用(SAR或不同分辨率的光学传感器)的挑战性数据进行评估。我们表明,我们的模型和算法在遥感特征提取和学习方面优于几种最先进的基线。
{"title":"Interactive Design of Object Classifiers in Remote Sensing","authors":"B. L. Saux","doi":"10.1109/ICPR.2014.444","DOIUrl":"https://doi.org/10.1109/ICPR.2014.444","url":null,"abstract":"This paper deals with the interactive design of generic classifiers for aerial images. In many real-life cases, object detectors that work are not available, due to a new geographical context or a need for a type of object unseen before. We propose an approach for on-line learning of such detectors using user interactions. Variants of gradient boosting and support-vector machine classification are proposed to cope with the problems raised by interactivity: unbalanced and partially mislabeled training data. We assess our framework for various visual classes (buildings, vegetation, cars, visual changes) on challenging data corresponding to several applications (SAR or optical sensors at various resolutions). We show that our model and algorithms outperform several state-of-the-art baselines for feature extraction and learning in remote sensing.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132361346","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Real-Time Visual Target Tracking in RGB-D Data for Person-Following Robots 基于RGB-D数据的人跟随机器人实时视觉目标跟踪
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.387
Youngwoo Yoon, Woo-han Yun, H. Yoon, Jaehong Kim
This paper describes a novel RGB-D-based visual target tracking method for person-following robots. We enhance a single-object tracker, which combines RGB and depth information, by exploiting two different types of distracters. First set of distracters includes objects existing near-by the target, and the other set is for objects looking similar to the target. The proposed algorithm reduces tracking drifts and wrong target re-identification by exploiting the distracters. Experiments on real-world video sequences demonstrating a person-following problem show a significant improvement over the method without tracking distracters and state-of-the-art RGB-based trackers. A mobile robot following a person is tested in real environment.
提出了一种新的基于rgb - d的人跟踪机器人视觉目标跟踪方法。我们通过利用两种不同类型的干扰来增强单目标跟踪器,该跟踪器结合了RGB和深度信息。第一组干扰物包括目标附近存在的物体,另一组是看起来与目标相似的物体。该算法利用干扰因素减少了跟踪漂移和错误目标的再识别。在真实世界的视频序列上进行的实验表明,与没有跟踪干扰物和最先进的基于rgb的跟踪器的方法相比,该方法有了显著的改进。在真实环境中对移动机器人跟随人进行了测试。
{"title":"Real-Time Visual Target Tracking in RGB-D Data for Person-Following Robots","authors":"Youngwoo Yoon, Woo-han Yun, H. Yoon, Jaehong Kim","doi":"10.1109/ICPR.2014.387","DOIUrl":"https://doi.org/10.1109/ICPR.2014.387","url":null,"abstract":"This paper describes a novel RGB-D-based visual target tracking method for person-following robots. We enhance a single-object tracker, which combines RGB and depth information, by exploiting two different types of distracters. First set of distracters includes objects existing near-by the target, and the other set is for objects looking similar to the target. The proposed algorithm reduces tracking drifts and wrong target re-identification by exploiting the distracters. Experiments on real-world video sequences demonstrating a person-following problem show a significant improvement over the method without tracking distracters and state-of-the-art RGB-based trackers. A mobile robot following a person is tested in real environment.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129576990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Vision-Based Road Bump Detection Using a Front-Mounted Car Camcorder 基于视觉的道路颠簸检测,使用前置车载摄像机
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.776
Hua-Tsung Chen, Chun-Yu Lai, Chun-Chieh Hsu, Suh-Yin Lee, B. Lin, Chien-Peng Ho
Advanced vehicle safety is a recently emerging issue, appealed from the rapidly explosive population of car owners. Increasing driver assistance systems have been designed for warning drivers of what should be noticed by analyzing surrounding environments with sensors and/or cameras. As one of the hazard road conditions, road bumps not only damage vehicles but also cause serious danger, especially at night or under poor lighting conditions. In this paper we propose a vision-based road bump detection system using a front-mounted car camcorder, which tends to be widespread deployed. First, the input video is transformed into a time-sliced image, which is a condensed video representation. Consequently, we estimate the vertical motion of the vehicle based on the time-sliced image and infer the existence of road bumps. Once a bump is detected, the location fix obtained from GPS is reported to a central server, so that the other vehicles can receive warnings when approaching the detected bumpy regions. Encouraging experimental results show that the proposed system can detect road bumps efficiently and effectively. It can be expected that traffic security will be significantly promoted through the mutually beneficial mechanism that a driver who is willing to report the bumps he/she meets can receive warnings issued from others as well.
先进的车辆安全是最近出现的一个问题,呼吁迅速爆炸的车主人口。越来越多的驾驶员辅助系统被设计用于通过传感器和/或摄像头分析周围环境来警告驾驶员应该注意什么。道路颠簸是危险的道路状况之一,不仅会损坏车辆,而且会造成严重的危险,特别是在夜间或光线不足的情况下。在本文中,我们提出了一种基于视觉的道路碰撞检测系统,该系统使用前置车载摄像机进行检测。首先,将输入视频转换为时间切片图像,这是一个压缩的视频表示。因此,我们根据时间切片图像估计车辆的垂直运动,并推断道路颠簸的存在。一旦检测到颠簸,从GPS获得的定位将报告给中央服务器,以便其他车辆在接近检测到的颠簸区域时可以收到警告。实验结果表明,该系统能够有效地检测道路颠簸。如果驾驶员愿意报告遇到的颠簸,也可以收到其他人发出的警告,这一互利机制将大大提高交通安全性。
{"title":"Vision-Based Road Bump Detection Using a Front-Mounted Car Camcorder","authors":"Hua-Tsung Chen, Chun-Yu Lai, Chun-Chieh Hsu, Suh-Yin Lee, B. Lin, Chien-Peng Ho","doi":"10.1109/ICPR.2014.776","DOIUrl":"https://doi.org/10.1109/ICPR.2014.776","url":null,"abstract":"Advanced vehicle safety is a recently emerging issue, appealed from the rapidly explosive population of car owners. Increasing driver assistance systems have been designed for warning drivers of what should be noticed by analyzing surrounding environments with sensors and/or cameras. As one of the hazard road conditions, road bumps not only damage vehicles but also cause serious danger, especially at night or under poor lighting conditions. In this paper we propose a vision-based road bump detection system using a front-mounted car camcorder, which tends to be widespread deployed. First, the input video is transformed into a time-sliced image, which is a condensed video representation. Consequently, we estimate the vertical motion of the vehicle based on the time-sliced image and infer the existence of road bumps. Once a bump is detected, the location fix obtained from GPS is reported to a central server, so that the other vehicles can receive warnings when approaching the detected bumpy regions. Encouraging experimental results show that the proposed system can detect road bumps efficiently and effectively. It can be expected that traffic security will be significantly promoted through the mutually beneficial mechanism that a driver who is willing to report the bumps he/she meets can receive warnings issued from others as well.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129207504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
An Improved Filtering for Fast Stereo Matching 一种改进的快速立体匹配滤波方法
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.423
Xiaoming Huang, Guoqin Cui, Yundong Zhang
This paper presents a novel full-image guided filtering based on eight-connected weight propagation for dense stereo matching. The proposed method has three main features: first, the proposed eight-connected weight propagation is more approximate compared to previous approach, second, the pixels employed into the filtering are all the pixels without constrained by one fixed window, last but not least, computational complexity of each pixel at each disparity level is 0(1), and the implementation of the filter can efficiently parallelized on hardware platform. Performance evaluation on Middlebury data sets shows that the proposed method is one of the best local algorithms in terms of both accuracy and speed.
提出了一种基于八连通权值传播的全像引导滤波方法,用于密集立体匹配。该方法具有三个主要特点:首先,所提出的八连通权值传播比以往的方法更接近;其次,用于滤波的像素都是不受一个固定窗口约束的像素;最后,每个视差级别上每个像素的计算复杂度为0(1),并且滤波器的实现可以在硬件平台上高效并行化。对Middlebury数据集的性能评估表明,该方法在精度和速度方面都是最好的局部算法之一。
{"title":"An Improved Filtering for Fast Stereo Matching","authors":"Xiaoming Huang, Guoqin Cui, Yundong Zhang","doi":"10.1109/ICPR.2014.423","DOIUrl":"https://doi.org/10.1109/ICPR.2014.423","url":null,"abstract":"This paper presents a novel full-image guided filtering based on eight-connected weight propagation for dense stereo matching. The proposed method has three main features: first, the proposed eight-connected weight propagation is more approximate compared to previous approach, second, the pixels employed into the filtering are all the pixels without constrained by one fixed window, last but not least, computational complexity of each pixel at each disparity level is 0(1), and the implementation of the filter can efficiently parallelized on hardware platform. Performance evaluation on Middlebury data sets shows that the proposed method is one of the best local algorithms in terms of both accuracy and speed.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125249711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Volume Reconstruction for MRI MRI体积重建
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.577
Meiqing Zhang, Huirao Nie, Yang Pei, L. Tao
One of the challenges in medical imaging is to increase the resolution of 3D MRI (Magnetic Resonance Imaging) signals. This is the problem of 3D signal reconstruction under the condition of very low sampling rate. Based on compressive sensing theory, the Direct Volume Reconstruction (DVR) method is proposed to reconstruct the 3D signal volume-by-volume based on a learned dictionary. DVR is a general method and applicable to any 3D signal as long as it can be sparsely represented. To exploit the nature of the 3D MRI system, the Progressive Volume Reconstruction (PVR) method is further proposed to improve the DVR reconstruction. In PVR, local reconstruction is used to reconstruct in-plane slices, and the output is then forwarded to global reconstruction, in which both the initially sampled and locally reconstructed signals are used together to reconstruct the whole 3D signal. Two separate dictionaries, rather than one, are trained in PVR. In this way, more prior knowledge from the training data is exploited. Experiments on a head MRI dataset demonstrate that DVR achieves much better performance than conventional tricubic interpolation and that PVR considerably improves DVR performance with regard to both PSNR and visibility quality.
提高三维磁共振成像(MRI)信号的分辨率是医学成像面临的挑战之一。这就是低采样率条件下的三维信号重建问题。基于压缩感知理论,提出了基于学习字典的直接体重构(DVR)方法对三维信号进行逐体重构。DVR是一种通用的方法,适用于任何三维信号,只要它能稀疏表示。针对三维MRI系统的特点,进一步提出渐进式体积重建(Progressive Volume Reconstruction, PVR)方法对DVR重建进行改进。在PVR中,通过局部重构对面内切片进行重构,然后将输出转发到全局重构,将初始采样的信号和局部重构的信号一起进行整体三维信号的重构。在PVR中训练两个独立的字典,而不是一个。通过这种方式,可以利用训练数据中的更多先验知识。在头部MRI数据集上的实验表明,DVR比传统的三次插值获得了更好的性能,并且PVR在PSNR和可见性质量方面都大大提高了DVR的性能。
{"title":"Volume Reconstruction for MRI","authors":"Meiqing Zhang, Huirao Nie, Yang Pei, L. Tao","doi":"10.1109/ICPR.2014.577","DOIUrl":"https://doi.org/10.1109/ICPR.2014.577","url":null,"abstract":"One of the challenges in medical imaging is to increase the resolution of 3D MRI (Magnetic Resonance Imaging) signals. This is the problem of 3D signal reconstruction under the condition of very low sampling rate. Based on compressive sensing theory, the Direct Volume Reconstruction (DVR) method is proposed to reconstruct the 3D signal volume-by-volume based on a learned dictionary. DVR is a general method and applicable to any 3D signal as long as it can be sparsely represented. To exploit the nature of the 3D MRI system, the Progressive Volume Reconstruction (PVR) method is further proposed to improve the DVR reconstruction. In PVR, local reconstruction is used to reconstruct in-plane slices, and the output is then forwarded to global reconstruction, in which both the initially sampled and locally reconstructed signals are used together to reconstruct the whole 3D signal. Two separate dictionaries, rather than one, are trained in PVR. In this way, more prior knowledge from the training data is exploited. Experiments on a head MRI dataset demonstrate that DVR achieves much better performance than conventional tricubic interpolation and that PVR considerably improves DVR performance with regard to both PSNR and visibility quality.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114263796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Temporal Mapping of Surveillance Video 监控视频的时间映射
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.707
S. Bagheri, J. Zheng
This paper visualizes surveillance video contents effectively in a 2D temporal image for indexing and retrieving a large video database. As an intermediate representation, this work extracts a temporal profile from video to convey accurate temporal information while keeps certain spatial characteristics for recognition. Different from spatial video indexing such as video synopsis that montages video frames for video summarization, our temporal indexing is achieved by designing a scheme to observe video from a side of video volume. A series of sampling lines are located over the field of view along the principal directions of targets to probe target motion, and the position varied temporal slices are obtained from the video volume. These slices are blended into the temporal profile with transparencies related to their spatial locations in the video. Static background and targets with less motion are further embedded into the temporal profile. Our goal is to provide less deformed shapes, more spatial information, and precise event occurrence in the temporal profile for video index and browsing.
本文将监控视频内容有效地可视化为二维时间图像,用于索引和检索大型视频数据库。作为一种中间表示,本作品从视频中提取时间轮廓,在传递准确的时间信息的同时,保留一定的空间特征以供识别。与空间视频索引(如视频摘要)通过蒙太奇视频帧进行视频摘要不同,我们的时间索引是通过设计一种从视频卷的侧面观察视频的方案来实现的。在视场上沿目标主方向设置一系列采样线,探测目标运动,从视频体中获得位置变化的时间切片。这些切片被混合到时间剖面中,其透明度与其在视频中的空间位置相关。静态背景和运动较少的目标进一步嵌入到时间轮廓中。我们的目标是为视频索引和浏览提供更少的变形形状,更多的空间信息和精确的事件发生时间剖面。
{"title":"Temporal Mapping of Surveillance Video","authors":"S. Bagheri, J. Zheng","doi":"10.1109/ICPR.2014.707","DOIUrl":"https://doi.org/10.1109/ICPR.2014.707","url":null,"abstract":"This paper visualizes surveillance video contents effectively in a 2D temporal image for indexing and retrieving a large video database. As an intermediate representation, this work extracts a temporal profile from video to convey accurate temporal information while keeps certain spatial characteristics for recognition. Different from spatial video indexing such as video synopsis that montages video frames for video summarization, our temporal indexing is achieved by designing a scheme to observe video from a side of video volume. A series of sampling lines are located over the field of view along the principal directions of targets to probe target motion, and the position varied temporal slices are obtained from the video volume. These slices are blended into the temporal profile with transparencies related to their spatial locations in the video. Static background and targets with less motion are further embedded into the temporal profile. Our goal is to provide less deformed shapes, more spatial information, and precise event occurrence in the temporal profile for video index and browsing.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114818053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Elastic Net Regularized Logistic Regression Using Cubic Majorization 基于三次优化的弹性网络正则化逻辑回归
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.593
M. Nilsson
In this work, a coordinate solver for elastic net regularized logistic regression is proposed. In particular, a method based on majorization maximization using a cubic function is derived. This to reliably and accurately optimize the objective function at each step without resorting to line search. Experiments show that the proposed solver is comparable to, or improves, state-of-the-art solvers. The proposed method is simpler, in the sense that there is no need for any line search, and can directly be used for small to large scale learning problems with elastic net regularization.
本文提出了弹性网正则化逻辑回归的坐标求解器。特别提出了一种基于三次函数的最大化方法。这使得目标函数在每一步都能可靠、准确地优化,而不需要进行直线搜索。实验表明,所提出的求解器可与最先进的求解器相媲美或改进。该方法更简单,不需要任何直线搜索,可以直接用于弹性网络正则化的小到大规模学习问题。
{"title":"Elastic Net Regularized Logistic Regression Using Cubic Majorization","authors":"M. Nilsson","doi":"10.1109/ICPR.2014.593","DOIUrl":"https://doi.org/10.1109/ICPR.2014.593","url":null,"abstract":"In this work, a coordinate solver for elastic net regularized logistic regression is proposed. In particular, a method based on majorization maximization using a cubic function is derived. This to reliably and accurately optimize the objective function at each step without resorting to line search. Experiments show that the proposed solver is comparable to, or improves, state-of-the-art solvers. The proposed method is simpler, in the sense that there is no need for any line search, and can directly be used for small to large scale learning problems with elastic net regularization.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115516087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2014 22nd International Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1