首页 > 最新文献

2016 IEEE International Conference on Image Processing (ICIP)最新文献

英文 中文
Text detection based on convolutional neural networks with spatial pyramid pooling 基于空间金字塔池的卷积神经网络文本检测
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532514
Rui Zhu, Xiao-Jiao Mao, Qi-Hai Zhu, Ning Li, Yubin Yang
Text detection is a difficult task due to the significant diversity of the texts appearing in natural scene images. In this paper, we propose a novel text descriptor, SPP-net, extracted by equipping the Convolutional Neural Network (CNN) with spatial pyramid pooling. We first compute the feature maps from the original text lines without any cropping or warping, and then generate the fixed-size representations for text discrimination. Experimental results on the latest ICDAR 2011 and 2013 datasets have proven that the proposed descriptor outperforms the state-of-the-art methods by a noticeable margin on F-measure with its merit of incorporating multi-scale text information and its flexibility of describing text regions with different sizes and shapes.
由于自然场景图像中出现的文本具有显著的多样性,文本检测是一项艰巨的任务。在本文中,我们提出了一种新的文本描述符SPP-net,它通过卷积神经网络(CNN)的空间金字塔池来提取。我们首先从原始文本行计算特征映射,不进行任何裁剪或扭曲,然后生成固定大小的文本区分表示。在最新的ICDAR 2011和2013数据集上的实验结果证明,该描述符具有融合多尺度文本信息的优点以及描述不同大小和形状的文本区域的灵活性,在F-measure上明显优于最先进的方法。
{"title":"Text detection based on convolutional neural networks with spatial pyramid pooling","authors":"Rui Zhu, Xiao-Jiao Mao, Qi-Hai Zhu, Ning Li, Yubin Yang","doi":"10.1109/ICIP.2016.7532514","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532514","url":null,"abstract":"Text detection is a difficult task due to the significant diversity of the texts appearing in natural scene images. In this paper, we propose a novel text descriptor, SPP-net, extracted by equipping the Convolutional Neural Network (CNN) with spatial pyramid pooling. We first compute the feature maps from the original text lines without any cropping or warping, and then generate the fixed-size representations for text discrimination. Experimental results on the latest ICDAR 2011 and 2013 datasets have proven that the proposed descriptor outperforms the state-of-the-art methods by a noticeable margin on F-measure with its merit of incorporating multi-scale text information and its flexibility of describing text regions with different sizes and shapes.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"167 1","pages":"1032-1036"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80530858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Content-adaptive low rank regularization for image denoising 图像去噪的内容自适应低秩正则化
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532928
Hangfan Liu, Xinfeng Zhang, Ruiqin Xiong
Prior knowledge plays an important role in image denoising tasks. This paper utilizes the data of the input image to adaptively model the prior distribution. The proposed scheme is based on the observation that, for a natural image, a matrix consisted of its vectorized non-local similar patches is of low rank. We use a non-convex smooth surrogate for the low-rank regularization, and view the optimization problem from the empirical Bayesian perspective. In such framework, a parameter-free distribution prior is derived from the grouped non-local similar image contents. Experimental results show that the proposed approach is highly competitive with several state-of-art denoising methods in PSNR and visual quality.
先验知识在图像去噪任务中起着重要的作用。本文利用输入图像的数据对先验分布进行自适应建模。该方案是基于对自然图像的观察,即由其矢量化的非局部相似块组成的矩阵秩低。我们使用非凸光滑代理来进行低秩正则化,并从经验贝叶斯的角度来看待优化问题。在该框架中,由分组的非局部相似图像内容导出无参数分布先验。实验结果表明,该方法在PSNR和视觉质量方面与几种最先进的去噪方法具有很强的竞争力。
{"title":"Content-adaptive low rank regularization for image denoising","authors":"Hangfan Liu, Xinfeng Zhang, Ruiqin Xiong","doi":"10.1109/ICIP.2016.7532928","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532928","url":null,"abstract":"Prior knowledge plays an important role in image denoising tasks. This paper utilizes the data of the input image to adaptively model the prior distribution. The proposed scheme is based on the observation that, for a natural image, a matrix consisted of its vectorized non-local similar patches is of low rank. We use a non-convex smooth surrogate for the low-rank regularization, and view the optimization problem from the empirical Bayesian perspective. In such framework, a parameter-free distribution prior is derived from the grouped non-local similar image contents. Experimental results show that the proposed approach is highly competitive with several state-of-art denoising methods in PSNR and visual quality.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"25 1 1","pages":"3091-3095"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82690449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
High-speed railway rod-insulator detection using segment clustering and deformable part models 基于分段聚类和可变形部件模型的高速铁路线路绝缘子检测
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533081
Ye Han, Zhigang Liu, Dah-Jye Lee, Guinan Zhang, Miao Deng
Catenary system maintenance is an important task to the operation of a high-seed railway system. Currently, the inspection of damaged parts in the catenary system is performed manually, which is often slow and unreliable. This paper proposes a method to detect and locate the rod-insulators in the image taken from the high-speed railway catenary system. Sub-images containing bar-shaped devices such as cantilever, strut, rod, and pole are first extracted from the image. Rod-insulator is then recognized and detected from these bar-shaped sub-images by using deformable part models and latent SVM. Experimental results show that the proposed method is able to locate rod-insulators accurately from the catenary image for the subsequent detect inspection process. The robustness of this method ensures its performance in different imaging conditions.
接触网系统维护是高速铁路系统运行的一项重要任务。目前,接触网系统中损坏部件的检测是手工进行的,这往往是缓慢和不可靠的。提出了一种高速铁路接触网图像中绝缘子的检测与定位方法。首先从图像中提取包含条形装置(如悬臂、支柱、杆和杆)的子图像。然后利用可变形零件模型和潜在支持向量机从这些条形子图像中识别和检测棒绝缘子。实验结果表明,该方法能够准确地从接触网图像中定位出杆状绝缘子,为后续的检测检测提供依据。该方法的鲁棒性保证了其在不同成像条件下的性能。
{"title":"High-speed railway rod-insulator detection using segment clustering and deformable part models","authors":"Ye Han, Zhigang Liu, Dah-Jye Lee, Guinan Zhang, Miao Deng","doi":"10.1109/ICIP.2016.7533081","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533081","url":null,"abstract":"Catenary system maintenance is an important task to the operation of a high-seed railway system. Currently, the inspection of damaged parts in the catenary system is performed manually, which is often slow and unreliable. This paper proposes a method to detect and locate the rod-insulators in the image taken from the high-speed railway catenary system. Sub-images containing bar-shaped devices such as cantilever, strut, rod, and pole are first extracted from the image. Rod-insulator is then recognized and detected from these bar-shaped sub-images by using deformable part models and latent SVM. Experimental results show that the proposed method is able to locate rod-insulators accurately from the catenary image for the subsequent detect inspection process. The robustness of this method ensures its performance in different imaging conditions.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"12 1","pages":"3852-3856"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82722752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
A weighted total variation approach for the atlas-based reconstruction of brain MR data 基于图谱的脑MR数据重建的加权总变异方法
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533177
Mingli Zhang, Kuldeep Kumar, Christian Desrosiers
Compressed sensing is a powerful approach to reconstruct high-quality images using a small number of samples. This paper presents a novel compressed sensing method that uses a probabilistic atlas to impose spatial constraints on the reconstruction of brain magnetic resonance imaging (MRI) data. A weighted total variation (TV) model is proposed to characterize the spatial distribution of gradients in the brain, and incorporate this information in the reconstruction process. Experiments on T1-weighted MR images from the ABIDE dataset show our proposed method to outperform the standard uniform TV model, as well as state-of-the-art approaches, for low sampling rates and high noise levels.
压缩感知是一种利用少量样本重建高质量图像的有效方法。本文提出了一种新的压缩感知方法,该方法利用概率图谱对脑磁共振成像(MRI)数据的重构施加空间约束。提出了一种加权总变异(TV)模型来表征大脑梯度的空间分布,并将该信息纳入重建过程。对来自ABIDE数据集的t1加权MR图像的实验表明,我们提出的方法在低采样率和高噪声水平下优于标准均匀电视模型以及最先进的方法。
{"title":"A weighted total variation approach for the atlas-based reconstruction of brain MR data","authors":"Mingli Zhang, Kuldeep Kumar, Christian Desrosiers","doi":"10.1109/ICIP.2016.7533177","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533177","url":null,"abstract":"Compressed sensing is a powerful approach to reconstruct high-quality images using a small number of samples. This paper presents a novel compressed sensing method that uses a probabilistic atlas to impose spatial constraints on the reconstruction of brain magnetic resonance imaging (MRI) data. A weighted total variation (TV) model is proposed to characterize the spatial distribution of gradients in the brain, and incorporate this information in the reconstruction process. Experiments on T1-weighted MR images from the ABIDE dataset show our proposed method to outperform the standard uniform TV model, as well as state-of-the-art approaches, for low sampling rates and high noise levels.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"18 1","pages":"4329-4333"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89001823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
H.264 intra coding with transforms based on prediction inaccuracy modeling 基于预测误差建模的H.264帧内编码
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532785
Xun Cai, J. Lim
In intra video coding, intra frames are predicted with intra prediction and the prediction residual signal is encoded. In many transform-based video coding systems, intra prediction residuals are encoded with transforms. For example, the Discrete Cosine Transform (DCT) and the Asymmetric Discrete Sine Transform (ADST) are used for intra prediction residuals in many coding systems. In the recent work, a set of transforms based on prediction inaccuracy modeling (PIM) has been proposed. These transforms are developed based on the observation that much of the residual non-stationarity is due to the use of an inaccurate prediction parameter. These transforms are shown to be effective for non-stationarity that arises in directional intra prediction residuals. In this paper, we implement the transforms based on prediction inaccuracy modeling on the H.264 intra coding system. The proposed transform is used in hybrid with the ADST. We compare the performance of the hybrid transform with the ADST and show that a significant bit-rate reduction is obtained with the proposed transform.
在视频内编码中,利用内预测对内帧进行预测,并对预测残差信号进行编码。在许多基于变换的视频编码系统中,用变换对帧内预测残差进行编码。例如,在许多编码系统中,离散余弦变换(DCT)和非对称离散正弦变换(ADST)被用于预测帧内残差。在最近的工作中,提出了一组基于预测不准确性建模的变换。这些变换是基于这样的观察,即大部分剩余非平稳性是由于使用了不准确的预测参数。这些变换被证明是有效的非平稳性产生的方向内预测残差。本文在H.264帧内编码系统上实现了基于预测误差建模的变换。所提出的变换与ADST混合使用。我们将混合变换与ADST的性能进行了比较,结果表明混合变换能显著降低比特率。
{"title":"H.264 intra coding with transforms based on prediction inaccuracy modeling","authors":"Xun Cai, J. Lim","doi":"10.1109/ICIP.2016.7532785","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532785","url":null,"abstract":"In intra video coding, intra frames are predicted with intra prediction and the prediction residual signal is encoded. In many transform-based video coding systems, intra prediction residuals are encoded with transforms. For example, the Discrete Cosine Transform (DCT) and the Asymmetric Discrete Sine Transform (ADST) are used for intra prediction residuals in many coding systems. In the recent work, a set of transforms based on prediction inaccuracy modeling (PIM) has been proposed. These transforms are developed based on the observation that much of the residual non-stationarity is due to the use of an inaccurate prediction parameter. These transforms are shown to be effective for non-stationarity that arises in directional intra prediction residuals. In this paper, we implement the transforms based on prediction inaccuracy modeling on the H.264 intra coding system. The proposed transform is used in hybrid with the ADST. We compare the performance of the hybrid transform with the ADST and show that a significant bit-rate reduction is obtained with the proposed transform.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"16 1","pages":"2380-2384"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87048311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Characterizing distortions in first-person videos 描述第一人称视频中的扭曲
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532797
Chen Bai, A. Reibman
First-person videos (FPVs) captured by wearable cameras often contain heavy distortions, including motion blur, rolling shutter artifacts and rotation. Existing image and video quality estimators are inefficient for this type of video. We develop a method specifically to measure the distortions present in FPVs, without using a high quality reference video. Our local visual information (LVI) algorithm measures motion blur, and we combine homography estimation with line angle histogram to measure rolling shutter artifacts and rotation. Our experiments demonstrate that captured FPVs have dramatically different distortions compared to traditional source videos. We also show that LVI is responsive to motion blur, but insensitive to rotation and shear.
可穿戴相机拍摄的第一人称视频(fps)通常存在严重失真,包括运动模糊、滚动快门伪影和旋转。现有的图像和视频质量估计器对于这种类型的视频是低效的。我们开发了一种专门测量fpv中存在的失真的方法,而不使用高质量的参考视频。我们的局部视觉信息(LVI)算法测量运动模糊,并将单应性估计与线角直方图相结合来测量滚动快门伪影和旋转。我们的实验表明,与传统源视频相比,捕获的fpv具有显着不同的失真。我们还表明LVI对运动模糊有响应,但对旋转和剪切不敏感。
{"title":"Characterizing distortions in first-person videos","authors":"Chen Bai, A. Reibman","doi":"10.1109/ICIP.2016.7532797","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532797","url":null,"abstract":"First-person videos (FPVs) captured by wearable cameras often contain heavy distortions, including motion blur, rolling shutter artifacts and rotation. Existing image and video quality estimators are inefficient for this type of video. We develop a method specifically to measure the distortions present in FPVs, without using a high quality reference video. Our local visual information (LVI) algorithm measures motion blur, and we combine homography estimation with line angle histogram to measure rolling shutter artifacts and rotation. Our experiments demonstrate that captured FPVs have dramatically different distortions compared to traditional source videos. We also show that LVI is responsive to motion blur, but insensitive to rotation and shear.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"58 1","pages":"2440-2444"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86965780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
HEVC still image coding and high efficiency image file format HEVC静态图像编码和高效的图像文件格式
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532321
J. Lainema, M. Hannuksela, V. Vadakital, Emre B. Aksu
The High Efficiency Video Coding (HEVC) standard includes support for a large range of image representation formats and provides an excellent image compression capability. The High Efficiency Image File Format (HEIF) offers a convenient way to encapsulate HEVC coded images, image sequences and animations together with associated metadata into a single file. This paper discusses various features and functionalities of the HEIF file format and compares the compression efficiency of HEVC still image coding to that of JPEG 2000. According to the experimental results HEVC provides about 25% bitrate reduction compared to JPEG 2000, while keeping the same objective picture quality.
高效视频编码(HEVC)标准包括对大量图像表示格式的支持,并提供了出色的图像压缩能力。高效图像文件格式(HEIF)提供了一种方便的方式来封装HEVC编码的图像、图像序列和动画以及相关的元数据到一个文件中。本文讨论了HEVC文件格式的各种特性和功能,并比较了HEVC静态图像编码与JPEG 2000的压缩效率。实验结果表明,HEVC在保持客观图像质量不变的情况下,比特率比JPEG 2000降低了25%左右。
{"title":"HEVC still image coding and high efficiency image file format","authors":"J. Lainema, M. Hannuksela, V. Vadakital, Emre B. Aksu","doi":"10.1109/ICIP.2016.7532321","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532321","url":null,"abstract":"The High Efficiency Video Coding (HEVC) standard includes support for a large range of image representation formats and provides an excellent image compression capability. The High Efficiency Image File Format (HEIF) offers a convenient way to encapsulate HEVC coded images, image sequences and animations together with associated metadata into a single file. This paper discusses various features and functionalities of the HEIF file format and compares the compression efficiency of HEVC still image coding to that of JPEG 2000. According to the experimental results HEVC provides about 25% bitrate reduction compared to JPEG 2000, while keeping the same objective picture quality.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"2015 1","pages":"71-75"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87946528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
One class classification applied in facial image analysis 一类分类在人脸图像分析中的应用
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532637
V. Mygdalis, Alexandros Iosifidis, A. Tefas, I. Pitas
In this paper, we apply One-Class Classification methods in facial image analysis problems. We consider the cases where the available training data information originates from one class, or one of the available classes is of high importance. We propose a novel extension of the One-Class Extreme Learning Machines algorithm aiming at minimizing both the training error and the data dispersion and consider solutions that generate decision functions in the ELM space, as well as in ELM spaces of arbitrary dimensionality. We evaluate the performance in publicly available datasets. The proposed method compares favourably to other state-of-the-art choices.
本文将一类分类方法应用于人脸图像分析问题。我们考虑可用的训练数据信息来自一个类的情况,或者其中一个可用的类是非常重要的。我们提出了一类极限学习机算法的新扩展,旨在最小化训练误差和数据分散,并考虑在ELM空间以及任意维的ELM空间中生成决策函数的解决方案。我们在公开可用的数据集中评估性能。所提出的方法比其他最先进的选择更有优势。
{"title":"One class classification applied in facial image analysis","authors":"V. Mygdalis, Alexandros Iosifidis, A. Tefas, I. Pitas","doi":"10.1109/ICIP.2016.7532637","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532637","url":null,"abstract":"In this paper, we apply One-Class Classification methods in facial image analysis problems. We consider the cases where the available training data information originates from one class, or one of the available classes is of high importance. We propose a novel extension of the One-Class Extreme Learning Machines algorithm aiming at minimizing both the training error and the data dispersion and consider solutions that generate decision functions in the ELM space, as well as in ELM spaces of arbitrary dimensionality. We evaluate the performance in publicly available datasets. The proposed method compares favourably to other state-of-the-art choices.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"1 1","pages":"1644-1648"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89972185","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Effective color correction pipeline for a noisy image 有效的颜色校正管道的噪声图像
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533111
Kenta Takahashi, Yusuke Monno, Masayuki Tanaka, M. Okutomi
Color correction is an essential image processing operation that transforms a camera-dependent RGB color space to a standard color space, e.g., the XYZ or the sRGB color space. The color correction is typically performed by multiplying the camera RGB values by a color correction matrix, which often amplifies image noise. In this paper, we propose an effective color correction pipeline for a noisy image. The proposed pipeline consists of two parts; the color correction and denoising. In the color correction part, we utilize spatially varying color correction (SVCC) that adaptively calculates the color correction matrices for each local image block considering the noise effect. Although the SVCC can effectively suppress the noise amplification, the noise is still included in the color corrected image, where the noise levels spatially vary for each local block. In the denoising part, we propose an effective denoising framework for the color corrected image with spatially varying noise levels. Experimental results demonstrate that the proposed color correction pipeline outperforms existing algorithms for various noise levels.
色彩校正是一项基本的图像处理操作,它将依赖于相机的RGB色彩空间转换为标准色彩空间,例如XYZ或sRGB色彩空间。色彩校正通常通过将相机RGB值乘以色彩校正矩阵来执行,这通常会放大图像噪声。本文提出了一种有效的噪声图像色彩校正管道。拟议的管道由两部分组成;色彩校正和去噪。在色彩校正部分,我们利用空间变化色彩校正(SVCC),考虑噪声影响自适应计算每个局部图像块的色彩校正矩阵。虽然SVCC可以有效地抑制噪声放大,但噪声仍然包含在颜色校正后的图像中,其中每个局部块的噪声水平在空间上是不同的。在去噪部分,我们提出了一种有效的去噪框架,用于具有空间变化噪声水平的彩色校正图像。实验结果表明,所提出的颜色校正管道在各种噪声水平下都优于现有算法。
{"title":"Effective color correction pipeline for a noisy image","authors":"Kenta Takahashi, Yusuke Monno, Masayuki Tanaka, M. Okutomi","doi":"10.1109/ICIP.2016.7533111","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533111","url":null,"abstract":"Color correction is an essential image processing operation that transforms a camera-dependent RGB color space to a standard color space, e.g., the XYZ or the sRGB color space. The color correction is typically performed by multiplying the camera RGB values by a color correction matrix, which often amplifies image noise. In this paper, we propose an effective color correction pipeline for a noisy image. The proposed pipeline consists of two parts; the color correction and denoising. In the color correction part, we utilize spatially varying color correction (SVCC) that adaptively calculates the color correction matrices for each local image block considering the noise effect. Although the SVCC can effectively suppress the noise amplification, the noise is still included in the color corrected image, where the noise levels spatially vary for each local block. In the denoising part, we propose an effective denoising framework for the color corrected image with spatially varying noise levels. Experimental results demonstrate that the proposed color correction pipeline outperforms existing algorithms for various noise levels.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"92 1","pages":"4002-4006"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89966764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Adaptive block truncation coding image compression technique using optimized dot diffusion 基于优化点扩散的自适应块截断编码图像压缩技术
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532736
Yun-Fu Liu, Jing-Ming Guo, Yu Cheng
Block truncation coding (BTC) has been considered as a highly efficient compression technique for decades, but the blocking artifact is its main issue. The halftoning-based BTC has significantly eased this issue, yet an apparent impulse noise artifact is accompanied. In this study, an improved BTC, termed adaptive dot-diffused BTC (ADBTC), is proposed to further improve the visual quality. Also, this method provides an additional flexibility on the compression ratios determination in contrast to the former fixed and few number of configuration possibilities. As documented in the experimental results, the proposed method achieves the superior image quality regarding the five various objective IQA methods. As a result, it is a very competitive approach for the needs of both high frame rate and high-resolution image compression.
块截断编码(BTC)是一种高效的压缩技术,但块伪影是其主要问题。基于半色调的比特币显著缓解了这一问题,但也伴随着明显的脉冲噪声伪影。本研究提出一种改进的自适应点扩散BTC (adaptive dot- diffusion BTC, ADBTC),以进一步改善视觉品质。此外,这种方法提供了一个额外的灵活性,在压缩比的确定相比,以前的固定和少数的配置可能性。实验结果表明,在五种不同的客观IQA方法中,本文提出的方法获得了较好的图像质量。因此,对于高帧率和高分辨率图像压缩的需求,它是一种非常有竞争力的方法。
{"title":"Adaptive block truncation coding image compression technique using optimized dot diffusion","authors":"Yun-Fu Liu, Jing-Ming Guo, Yu Cheng","doi":"10.1109/ICIP.2016.7532736","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532736","url":null,"abstract":"Block truncation coding (BTC) has been considered as a highly efficient compression technique for decades, but the blocking artifact is its main issue. The halftoning-based BTC has significantly eased this issue, yet an apparent impulse noise artifact is accompanied. In this study, an improved BTC, termed adaptive dot-diffused BTC (ADBTC), is proposed to further improve the visual quality. Also, this method provides an additional flexibility on the compression ratios determination in contrast to the former fixed and few number of configuration possibilities. As documented in the experimental results, the proposed method achieves the superior image quality regarding the five various objective IQA methods. As a result, it is a very competitive approach for the needs of both high frame rate and high-resolution image compression.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"9 1","pages":"2137-2141"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82220282","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2016 IEEE International Conference on Image Processing (ICIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1