首页 > 最新文献

2016 IEEE International Conference on Image Processing (ICIP)最新文献

英文 中文
Effective color correction pipeline for a noisy image 有效的颜色校正管道的噪声图像
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533111
Kenta Takahashi, Yusuke Monno, Masayuki Tanaka, M. Okutomi
Color correction is an essential image processing operation that transforms a camera-dependent RGB color space to a standard color space, e.g., the XYZ or the sRGB color space. The color correction is typically performed by multiplying the camera RGB values by a color correction matrix, which often amplifies image noise. In this paper, we propose an effective color correction pipeline for a noisy image. The proposed pipeline consists of two parts; the color correction and denoising. In the color correction part, we utilize spatially varying color correction (SVCC) that adaptively calculates the color correction matrices for each local image block considering the noise effect. Although the SVCC can effectively suppress the noise amplification, the noise is still included in the color corrected image, where the noise levels spatially vary for each local block. In the denoising part, we propose an effective denoising framework for the color corrected image with spatially varying noise levels. Experimental results demonstrate that the proposed color correction pipeline outperforms existing algorithms for various noise levels.
色彩校正是一项基本的图像处理操作,它将依赖于相机的RGB色彩空间转换为标准色彩空间,例如XYZ或sRGB色彩空间。色彩校正通常通过将相机RGB值乘以色彩校正矩阵来执行,这通常会放大图像噪声。本文提出了一种有效的噪声图像色彩校正管道。拟议的管道由两部分组成;色彩校正和去噪。在色彩校正部分,我们利用空间变化色彩校正(SVCC),考虑噪声影响自适应计算每个局部图像块的色彩校正矩阵。虽然SVCC可以有效地抑制噪声放大,但噪声仍然包含在颜色校正后的图像中,其中每个局部块的噪声水平在空间上是不同的。在去噪部分,我们提出了一种有效的去噪框架,用于具有空间变化噪声水平的彩色校正图像。实验结果表明,所提出的颜色校正管道在各种噪声水平下都优于现有算法。
{"title":"Effective color correction pipeline for a noisy image","authors":"Kenta Takahashi, Yusuke Monno, Masayuki Tanaka, M. Okutomi","doi":"10.1109/ICIP.2016.7533111","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533111","url":null,"abstract":"Color correction is an essential image processing operation that transforms a camera-dependent RGB color space to a standard color space, e.g., the XYZ or the sRGB color space. The color correction is typically performed by multiplying the camera RGB values by a color correction matrix, which often amplifies image noise. In this paper, we propose an effective color correction pipeline for a noisy image. The proposed pipeline consists of two parts; the color correction and denoising. In the color correction part, we utilize spatially varying color correction (SVCC) that adaptively calculates the color correction matrices for each local image block considering the noise effect. Although the SVCC can effectively suppress the noise amplification, the noise is still included in the color corrected image, where the noise levels spatially vary for each local block. In the denoising part, we propose an effective denoising framework for the color corrected image with spatially varying noise levels. Experimental results demonstrate that the proposed color correction pipeline outperforms existing algorithms for various noise levels.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"92 1","pages":"4002-4006"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89966764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
One class classification applied in facial image analysis 一类分类在人脸图像分析中的应用
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532637
V. Mygdalis, Alexandros Iosifidis, A. Tefas, I. Pitas
In this paper, we apply One-Class Classification methods in facial image analysis problems. We consider the cases where the available training data information originates from one class, or one of the available classes is of high importance. We propose a novel extension of the One-Class Extreme Learning Machines algorithm aiming at minimizing both the training error and the data dispersion and consider solutions that generate decision functions in the ELM space, as well as in ELM spaces of arbitrary dimensionality. We evaluate the performance in publicly available datasets. The proposed method compares favourably to other state-of-the-art choices.
本文将一类分类方法应用于人脸图像分析问题。我们考虑可用的训练数据信息来自一个类的情况,或者其中一个可用的类是非常重要的。我们提出了一类极限学习机算法的新扩展,旨在最小化训练误差和数据分散,并考虑在ELM空间以及任意维的ELM空间中生成决策函数的解决方案。我们在公开可用的数据集中评估性能。所提出的方法比其他最先进的选择更有优势。
{"title":"One class classification applied in facial image analysis","authors":"V. Mygdalis, Alexandros Iosifidis, A. Tefas, I. Pitas","doi":"10.1109/ICIP.2016.7532637","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532637","url":null,"abstract":"In this paper, we apply One-Class Classification methods in facial image analysis problems. We consider the cases where the available training data information originates from one class, or one of the available classes is of high importance. We propose a novel extension of the One-Class Extreme Learning Machines algorithm aiming at minimizing both the training error and the data dispersion and consider solutions that generate decision functions in the ELM space, as well as in ELM spaces of arbitrary dimensionality. We evaluate the performance in publicly available datasets. The proposed method compares favourably to other state-of-the-art choices.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"1 1","pages":"1644-1648"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89972185","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
CNN-aware binary MAP for general semantic segmentation 用于一般语义分割的cnn感知二进制MAP
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532693
Mahdyar Ravanbakhsh, Hossein Mousavi, Moin Nabi, Mohammad Rastegari, C. Regazzoni
In this paper we introduce a novel method for general semantic segmentation that can benefit from general semantics of Convolutional Neural Network (CNN). Our segmentation proposes visually and semantically coherent image segments. We use binary encoding of CNN features to overcome the difficulty of the clustering on the high-dimensional CNN feature space. These binary codes are very robust against noise and non-semantic changes in the image. These binary encoding can be embedded into the CNN as an extra layer at the end of the network. This results in real-time segmentation. To the best of our knowledge our method is the first attempt on general semantic image segmentation using CNN. All the previous papers were limited to few number of category of the images (e.g. PASCAL VOC). Experiments show that our segmentation algorithm outperform the state-of-the-art non-semantic segmentation methods by large margin.
本文介绍了一种利用卷积神经网络(CNN)的通用语义进行通用语义分割的新方法。我们的分割提出了视觉上和语义上连贯的图像片段。我们使用CNN特征的二值编码来克服CNN高维特征空间上聚类的困难。这些二进制代码对图像中的噪声和非语义变化具有很强的鲁棒性。这些二进制编码可以作为网络末端的额外层嵌入到CNN中。这导致了实时分割。据我们所知,我们的方法是第一次尝试使用CNN进行一般语义图像分割。以往的所有论文都局限于图像的少数几个类别(如PASCAL VOC)。实验表明,我们的分割算法在很大程度上优于目前最先进的非语义分割方法。
{"title":"CNN-aware binary MAP for general semantic segmentation","authors":"Mahdyar Ravanbakhsh, Hossein Mousavi, Moin Nabi, Mohammad Rastegari, C. Regazzoni","doi":"10.1109/ICIP.2016.7532693","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532693","url":null,"abstract":"In this paper we introduce a novel method for general semantic segmentation that can benefit from general semantics of Convolutional Neural Network (CNN). Our segmentation proposes visually and semantically coherent image segments. We use binary encoding of CNN features to overcome the difficulty of the clustering on the high-dimensional CNN feature space. These binary codes are very robust against noise and non-semantic changes in the image. These binary encoding can be embedded into the CNN as an extra layer at the end of the network. This results in real-time segmentation. To the best of our knowledge our method is the first attempt on general semantic image segmentation using CNN. All the previous papers were limited to few number of category of the images (e.g. PASCAL VOC). Experiments show that our segmentation algorithm outperform the state-of-the-art non-semantic segmentation methods by large margin.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"63 1","pages":"1923-1927"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91001204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Segmentation and classification of melanocytic skin lesions using local and contextual features 使用局部和上下文特征的黑素细胞性皮肤病变的分割和分类
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532836
Eliezer Bernart, J. Scharcanski, S. Bampi
This work presents a novel approach for detecting and classifying melanocytic skin lesions on macroscopic images. We oversegment the skin lesions using superpixels, and classify independently each superpixel as a benign or malignant using local and contextual information. The overall superpixel classification results allow to calculate an index of malignancy or benignity for the skin lesion. Using the proposed approach it is possible to discriminate a malignant from a benign skin lesion by recognizing early signs of malignancy in parts of the segmented skin lesion. The experimental results are promising, and show a potential accuracy of 99.34% on a popular data set, outperforming the current state-of-art methods.
这项工作提出了一种新的方法来检测和分类黑素细胞皮肤病变的宏观图像。我们使用超像素对皮肤病变进行过分割,并使用局部和上下文信息独立地将每个超像素分类为良性或恶性。总体超像素分类结果允许计算恶性或良性皮肤病变的指数。使用所提出的方法,可以通过识别部分分段皮肤病变的早期恶性体征来区分恶性皮肤病变和良性皮肤病变。实验结果很有希望,在一个流行的数据集上显示出99.34%的潜在准确率,优于目前最先进的方法。
{"title":"Segmentation and classification of melanocytic skin lesions using local and contextual features","authors":"Eliezer Bernart, J. Scharcanski, S. Bampi","doi":"10.1109/ICIP.2016.7532836","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532836","url":null,"abstract":"This work presents a novel approach for detecting and classifying melanocytic skin lesions on macroscopic images. We oversegment the skin lesions using superpixels, and classify independently each superpixel as a benign or malignant using local and contextual information. The overall superpixel classification results allow to calculate an index of malignancy or benignity for the skin lesion. Using the proposed approach it is possible to discriminate a malignant from a benign skin lesion by recognizing early signs of malignancy in parts of the segmented skin lesion. The experimental results are promising, and show a potential accuracy of 99.34% on a popular data set, outperforming the current state-of-art methods.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"19 16","pages":"2633-2637"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91446880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Quality assessment of 3D synthesized images via disoccluded region discovery 利用去闭塞区发现技术评价三维合成图像的质量
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532510
Yu Zhou, Leida Li, Ke Gu, Yuming Fang, Weisi Lin
Depth-Image-Based-Rendering (DIBR) is fundamental in free-viewpoint 3D video, which has been widely used to generate synthesized views from multi-view images. The majority of DIBR algorithms cause disoccluded regions, which are the areas invisible in original views but emerge in synthesized views. The quality of synthesized images is mainly contaminated by distortions in these disoccluded regions. Unfortunately, traditional image quality metrics are not effective for these synthesized images because they are sensitive to geometric distortions. To solve the problem, this paper proposes an objective quality evaluation method for 3D Synthesized images via Disoccluded Region Discovery (SDRD). A self-adaptive scale transform model is first adopted to preprocess the images on account of the impacts of view distance. Then disoccluded regions are detected by comparing the absolute difference between the preprocessed synthesized image and the warped image of preprocessed reference image. Furthermore, the disoccluded regions are weighted by a weighting function proposed to account for the varying sensitivities of human eyes to the size of disoccluded regions. Experiments conducted on IRCCyN/IVC DIBR image database demonstrate that the proposed SDRD method remarkably outperforms traditional 2D and existing DIBR-related quality metrics.
深度图像渲染(deep - image - based rendering, DIBR)是自由视点三维视频的基础,它被广泛用于从多视点图像生成合成视图。大多数DIBR算法都会产生闭塞区域,即在原始视图中不可见但在合成视图中出现的区域。合成图像的质量主要受到这些去遮挡区域畸变的影响。遗憾的是,传统的图像质量指标对这些合成图像并不有效,因为它们对几何畸变很敏感。为了解决这一问题,本文提出了一种基于去遮挡区域发现(Disoccluded Region Discovery, SDRD)的三维合成图像质量客观评价方法。考虑到视距的影响,首先采用自适应尺度变换模型对图像进行预处理。然后通过比较预处理后的合成图像与预处理后的参考图像的扭曲图像的绝对差值来检测去闭塞区域。此外,利用加权函数对解除遮挡区域进行加权,以考虑人眼对解除遮挡区域大小的不同敏感性。在IRCCyN/IVC DIBR图像数据库上进行的实验表明,提出的SDRD方法显著优于传统的2D和现有的DIBR相关质量指标。
{"title":"Quality assessment of 3D synthesized images via disoccluded region discovery","authors":"Yu Zhou, Leida Li, Ke Gu, Yuming Fang, Weisi Lin","doi":"10.1109/ICIP.2016.7532510","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532510","url":null,"abstract":"Depth-Image-Based-Rendering (DIBR) is fundamental in free-viewpoint 3D video, which has been widely used to generate synthesized views from multi-view images. The majority of DIBR algorithms cause disoccluded regions, which are the areas invisible in original views but emerge in synthesized views. The quality of synthesized images is mainly contaminated by distortions in these disoccluded regions. Unfortunately, traditional image quality metrics are not effective for these synthesized images because they are sensitive to geometric distortions. To solve the problem, this paper proposes an objective quality evaluation method for 3D Synthesized images via Disoccluded Region Discovery (SDRD). A self-adaptive scale transform model is first adopted to preprocess the images on account of the impacts of view distance. Then disoccluded regions are detected by comparing the absolute difference between the preprocessed synthesized image and the warped image of preprocessed reference image. Furthermore, the disoccluded regions are weighted by a weighting function proposed to account for the varying sensitivities of human eyes to the size of disoccluded regions. Experiments conducted on IRCCyN/IVC DIBR image database demonstrate that the proposed SDRD method remarkably outperforms traditional 2D and existing DIBR-related quality metrics.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"12 1","pages":"1012-1016"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84817468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
A perceptual visibility metric for banding artifacts 条带伪影的感知可见性度量
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532722
Yilin Wang, S. Kum, Chao Chen, A. Kokaram
Banding is a common video artifact caused by compressing low texture regions with coarse quantization. Relatively few previous attempts exist to address banding and none incorporate subjective testing for calibrating the measurement. In this paper, we propose a novel metric that incorporates both edge length and contrast across the edge to measure video banding. We further introduce both reference and non-reference metrics. Our results demonstrate that the new metrics have a very high correlation with subjective assessment and certainly outperforms PSNR, SSIM, and VQM.
条带是对低纹理区域进行粗量化压缩后产生的常见视频伪影。相对较少的先前的尝试存在,以解决分带和没有纳入主观测试校准测量。在本文中,我们提出了一种结合边缘长度和边缘对比度的新度量来测量视频带。我们进一步介绍参考指标和非参考指标。我们的结果表明,新的指标与主观评估有非常高的相关性,并且肯定优于PSNR、SSIM和VQM。
{"title":"A perceptual visibility metric for banding artifacts","authors":"Yilin Wang, S. Kum, Chao Chen, A. Kokaram","doi":"10.1109/ICIP.2016.7532722","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532722","url":null,"abstract":"Banding is a common video artifact caused by compressing low texture regions with coarse quantization. Relatively few previous attempts exist to address banding and none incorporate subjective testing for calibrating the measurement. In this paper, we propose a novel metric that incorporates both edge length and contrast across the edge to measure video banding. We further introduce both reference and non-reference metrics. Our results demonstrate that the new metrics have a very high correlation with subjective assessment and certainly outperforms PSNR, SSIM, and VQM.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"31 1","pages":"2067-2071"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77183986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Unsupervised person re-identification with locality-constrained Earth Mover's distance 无监督人员再识别与位置约束的土移者的距离
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533169
Dan Wang, Canxiang Yan, S. Shan, Xilin Chen
The difficult acquisition of labeled data and the misalignment of local matching are major obstacles to apply person re-identification in real scenarios. To alleviate these problems, we propose an unsupervised method, called locality-constrained Earth Mover's Distance (LC-EMD), to learn the optimal measure between image pairs. Specifically, Gaussian mixture models (GMMs) are learned as signatures. By imposing locality constraints, LC-EMD can naturally achieve partial matching between Gaussian components. Moreover, LC-EMD has the analytical solution which can be efficiently computed. Experiments on two public datasets demonstrate LC-EMD is robust to misalignment and performs better than other unsupervised methods.
标记数据的获取困难和局部匹配的不匹配是在真实场景中应用人员再识别的主要障碍。为了解决这些问题,我们提出了一种无监督的方法,称为位置约束的地球移动者距离(LC-EMD),以学习图像对之间的最佳度量。具体来说,高斯混合模型(gmm)作为签名学习。通过施加局部性约束,LC-EMD可以很自然地实现高斯分量之间的部分匹配。此外,LC-EMD具有解析解,可以有效地进行计算。在两个公共数据集上的实验表明,LC-EMD对不对准具有鲁棒性,并且优于其他无监督方法。
{"title":"Unsupervised person re-identification with locality-constrained Earth Mover's distance","authors":"Dan Wang, Canxiang Yan, S. Shan, Xilin Chen","doi":"10.1109/ICIP.2016.7533169","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533169","url":null,"abstract":"The difficult acquisition of labeled data and the misalignment of local matching are major obstacles to apply person re-identification in real scenarios. To alleviate these problems, we propose an unsupervised method, called locality-constrained Earth Mover's Distance (LC-EMD), to learn the optimal measure between image pairs. Specifically, Gaussian mixture models (GMMs) are learned as signatures. By imposing locality constraints, LC-EMD can naturally achieve partial matching between Gaussian components. Moreover, LC-EMD has the analytical solution which can be efficiently computed. Experiments on two public datasets demonstrate LC-EMD is robust to misalignment and performs better than other unsupervised methods.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"115 1","pages":"4289-4293"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77926931","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Learning a multiscale patch-based representation for image denoising in X-RAY fluoroscopy 学习基于多尺度补丁的x射线透视图像去噪方法
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532775
Y. Matviychuk, B. Mailhé, Xiao Chen, Qiu Wang, A. Kiraly, N. Strobel, M. Nadar
Denoising is an indispensable step in processing low-dose X-ray fluoroscopic images that requires development of specialized high-quality algorithms able to operate in near real-time. We address this problem with an efficient deep learning approach based on the process-centric view of traditional iterative thresholding methods. We develop a novel trainable patch-based multiscale framework for sparse image representation. In a computationally efficient way, it allows us to accurately reconstruct important image features on multiple levels of decomposition with patch dictionaries of reduced size and complexity. The flexibility of the chosen machine learning approach allows us to tailor the learned basis for preserving important structural information in the image and noticeably minimize the amount of artifacts. Our denoising results obtained with real clinical data demonstrate significant quality improvement and are computed much faster in comparison with the BM3D algorithm.
去噪是处理低剂量x射线透视图像必不可少的步骤,这需要开发能够近实时操作的专门高质量算法。我们用一种高效的深度学习方法来解决这个问题,这种方法基于传统迭代阈值方法的以过程为中心的观点。我们开发了一种新的基于可训练补丁的多尺度稀疏图像表示框架。在一个计算效率高的方式,它允许我们准确地重建重要的图像特征在多层次的分解与补丁字典的大小和复杂性减少。所选择的机器学习方法的灵活性使我们能够定制学习基础,以保留图像中的重要结构信息,并显着减少人工制品的数量。与BM3D算法相比,我们对真实临床数据的去噪结果显示出明显的质量改善,并且计算速度更快。
{"title":"Learning a multiscale patch-based representation for image denoising in X-RAY fluoroscopy","authors":"Y. Matviychuk, B. Mailhé, Xiao Chen, Qiu Wang, A. Kiraly, N. Strobel, M. Nadar","doi":"10.1109/ICIP.2016.7532775","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532775","url":null,"abstract":"Denoising is an indispensable step in processing low-dose X-ray fluoroscopic images that requires development of specialized high-quality algorithms able to operate in near real-time. We address this problem with an efficient deep learning approach based on the process-centric view of traditional iterative thresholding methods. We develop a novel trainable patch-based multiscale framework for sparse image representation. In a computationally efficient way, it allows us to accurately reconstruct important image features on multiple levels of decomposition with patch dictionaries of reduced size and complexity. The flexibility of the chosen machine learning approach allows us to tailor the learned basis for preserving important structural information in the image and noticeably minimize the amount of artifacts. Our denoising results obtained with real clinical data demonstrate significant quality improvement and are computed much faster in comparison with the BM3D algorithm.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"6 2 1","pages":"2330-2334"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78370389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Person re-identification based on hierarchical bipartite graph matching 基于层次二部图匹配的人物再识别
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533162
Yan Huang, Hao Sheng, Z. Xiong
This work proposes a novel person re-identification method based on Hierarchical Bipartite Graph Matching. Because human eyes observe person appearance roughly first and then goes further into the details gradually, our method abstracts person image from coarse to fine granularity, and finally into a three layer tree structure. Then, three bipartite graph matching methods are proposed for the matching of each layer between the trees. At the bottom layer Non-complete Bipartite Graph matching is proposed to collect matching pairs among small local regions. At the middle layer Semi-complete Bipartite Graph matching is used to deal with the problem of spatial misalignment between two person bodies. Complete Bipartite Graph matching is presented to refine the ranking result at the top layer. The effectiveness of our method is validated on the CAVIAR4REID and VIPeR datasets, and competitive results are achieved on both datasets.
本文提出了一种基于层次二部图匹配的人物再识别方法。由于人眼先对人的外表进行粗略的观察,然后逐渐深入到细节,因此我们的方法将人的图像从粗粒度抽象到细粒度,最后形成三层树状结构。然后,提出了三种二部图匹配方法,用于树间各层的匹配。在底层提出非完全二部图匹配,收集小局部区域之间的匹配对。在中间层,采用半完全二部图匹配来处理两个人体之间的空间不对齐问题。采用完全二部图匹配的方法对顶层的排序结果进行细化。在CAVIAR4REID和VIPeR数据集上验证了该方法的有效性,并在两个数据集上取得了具有竞争力的结果。
{"title":"Person re-identification based on hierarchical bipartite graph matching","authors":"Yan Huang, Hao Sheng, Z. Xiong","doi":"10.1109/ICIP.2016.7533162","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533162","url":null,"abstract":"This work proposes a novel person re-identification method based on Hierarchical Bipartite Graph Matching. Because human eyes observe person appearance roughly first and then goes further into the details gradually, our method abstracts person image from coarse to fine granularity, and finally into a three layer tree structure. Then, three bipartite graph matching methods are proposed for the matching of each layer between the trees. At the bottom layer Non-complete Bipartite Graph matching is proposed to collect matching pairs among small local regions. At the middle layer Semi-complete Bipartite Graph matching is used to deal with the problem of spatial misalignment between two person bodies. Complete Bipartite Graph matching is presented to refine the ranking result at the top layer. The effectiveness of our method is validated on the CAVIAR4REID and VIPeR datasets, and competitive results are achieved on both datasets.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"13 1","pages":"4255-4259"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78841906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Reduction of false alarms triggered by spiders/cobwebs in surveillance camera networks 减少监控摄像机网络中蜘蛛/蜘蛛网引发的误报
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532496
R. Hebbalaguppe, Kevin McGuinness, J. Kuklyte, Rami Albatal, C. Direkoğlu, N. O’Connor
The percentage of false alarms caused by spiders in automated surveillance can range from 20-50%. False alarms increase the workload of surveillance personnel validating the alarms and the maintenance labor cost associated with regular cleaning of webs. We propose a novel, cost effective method to detect false alarms triggered by spiders/webs in surveillance camera networks. This is accomplished by building a spider classifier intended to be a part of the surveillance video processing pipeline. The proposed method uses a feature descriptor obtained by early fusion of blur and texture. The approach is sufficiently efficient for real-time processing and yet comparable in performance with more computationally costly approaches like SIFT with bag of visual words aggregation. The proposed method can eliminate 98.5% of false alarms caused by spiders in a data set supplied by an industry partner, with a false positive rate of less than 1%.
在自动监控中,蜘蛛引起的误报百分比可以在20-50%之间。虚警增加了监控人员验证虚警的工作量,也增加了定期清理网络的维护人力成本。我们提出了一种新颖的、经济有效的方法来检测监控摄像机网络中蜘蛛/网触发的假警报。这是通过构建一个蜘蛛分类器来实现的,该分类器旨在成为监控视频处理管道的一部分。该方法利用模糊和纹理的早期融合得到的特征描述符。该方法对于实时处理足够有效,但在性能上可与计算成本更高的方法相媲美,例如具有视觉单词聚合包的SIFT。在行业合作伙伴提供的数据集中,提出的方法可以消除98.5%的蜘蛛引起的误报,误报率小于1%。
{"title":"Reduction of false alarms triggered by spiders/cobwebs in surveillance camera networks","authors":"R. Hebbalaguppe, Kevin McGuinness, J. Kuklyte, Rami Albatal, C. Direkoğlu, N. O’Connor","doi":"10.1109/ICIP.2016.7532496","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532496","url":null,"abstract":"The percentage of false alarms caused by spiders in automated surveillance can range from 20-50%. False alarms increase the workload of surveillance personnel validating the alarms and the maintenance labor cost associated with regular cleaning of webs. We propose a novel, cost effective method to detect false alarms triggered by spiders/webs in surveillance camera networks. This is accomplished by building a spider classifier intended to be a part of the surveillance video processing pipeline. The proposed method uses a feature descriptor obtained by early fusion of blur and texture. The approach is sufficiently efficient for real-time processing and yet comparable in performance with more computationally costly approaches like SIFT with bag of visual words aggregation. The proposed method can eliminate 98.5% of false alarms caused by spiders in a data set supplied by an industry partner, with a false positive rate of less than 1%.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"1 1","pages":"943-947"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79484024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2016 IEEE International Conference on Image Processing (ICIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1