首页 > 最新文献

2016 IEEE International Conference on Image Processing (ICIP)最新文献

英文 中文
Block-size adaptive transform domain estimation of end-to-end distortion for error-resilient video coding 纠错视频编码端到端失真的块大小自适应变换域估计
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532727
Bohan Li, Tejaswi Nanjundaswamy, K. Rose
The accuracy of end-to-end distortion (EED) estimation is crucial to achieving effective error resilient video coding. An established solution, the recursive optimal per-pixel estimate (ROPE), does so by tracking the first and second moments of decoder-reconstructed pixels. An alternative estimation approach, the spectral coefficient-wise optimal recursive estimate (SCORE), tracks instead moments of decoder-reconstructed transform coefficients, which enables accounting for transform domain operations. However, the SCORE formulation relies on a fixed transform block size, which is incompatible with recent standards. This paper proposes a non-trivial generalization of the SCORE framework which, in particular, accounts for arbitrary block size combinations involving the current and reference block partitions. This seemingly intractable objective is achieved by a two-step approach: i) Given the fixed block size moments of a reference frame, estimate moments of transform coefficients for the codec-selected current block partition; ii) Convert the current results to transform coefficient moments corresponding to a regular fixed block size grid, to facilitate EED estimation for the next frame. Experimental results first demonstrate the accuracy of the proposed estimate in conjunction with transform domain temporal prediction. Then the estimate is leveraged to optimize the coding mode and yields considerable gains in rate-distortion performance.
端到端失真估计的准确性是实现有效的抗错误性视频编码的关键。一个已建立的解决方案,递归最优逐像素估计(ROPE),通过跟踪解码器重构像素的第一和第二矩来实现。另一种估计方法,谱系数最优递归估计(SCORE),跟踪解码器重构变换系数的矩,从而可以考虑变换域操作。然而,SCORE公式依赖于固定的转换块大小,这与最近的标准不兼容。本文提出了SCORE框架的非平凡泛化,特别是考虑到涉及当前和参考块分区的任意块大小组合。这个看似棘手的目标是通过两步方法实现的:i)给定参考帧的固定块大小矩,估计编解码器选择的当前块分割的变换系数矩;ii)将当前结果转换为规则的固定块大小网格对应的变换系数矩,便于下一帧的EED估计。实验结果首先证明了该方法与变换域时间预测相结合的准确性。然后利用估计来优化编码模式,并在率失真性能方面获得可观的收益。
{"title":"Block-size adaptive transform domain estimation of end-to-end distortion for error-resilient video coding","authors":"Bohan Li, Tejaswi Nanjundaswamy, K. Rose","doi":"10.1109/ICIP.2016.7532727","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532727","url":null,"abstract":"The accuracy of end-to-end distortion (EED) estimation is crucial to achieving effective error resilient video coding. An established solution, the recursive optimal per-pixel estimate (ROPE), does so by tracking the first and second moments of decoder-reconstructed pixels. An alternative estimation approach, the spectral coefficient-wise optimal recursive estimate (SCORE), tracks instead moments of decoder-reconstructed transform coefficients, which enables accounting for transform domain operations. However, the SCORE formulation relies on a fixed transform block size, which is incompatible with recent standards. This paper proposes a non-trivial generalization of the SCORE framework which, in particular, accounts for arbitrary block size combinations involving the current and reference block partitions. This seemingly intractable objective is achieved by a two-step approach: i) Given the fixed block size moments of a reference frame, estimate moments of transform coefficients for the codec-selected current block partition; ii) Convert the current results to transform coefficient moments corresponding to a regular fixed block size grid, to facilitate EED estimation for the next frame. Experimental results first demonstrate the accuracy of the proposed estimate in conjunction with transform domain temporal prediction. Then the estimate is leveraged to optimize the coding mode and yields considerable gains in rate-distortion performance.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"7 1","pages":"2092-2096"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82138938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Material segmentation in hyperspectral images with minimal region perimeters 最小区域周长高光谱图像的材料分割
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532474
Yu Zhang, C. P. Huynh, N. Habili, K. Ngan
We propose a supervised approach to the classification and segmentation of material regions in hyperspectral imagery. Our algorithm is a two-stage process, combining a pixelwise classification step with a segmentation step aiming to minimise the total perimeters of the resulting regions. Our algorithm is distinctive in its ability to ensure label consistency within local homogeneous areas and to generate material segments with smooth boundaries. Furthermore, we establish a new hyperspectral benchmark dataset to demonstrate the advantages of the proposed approach over several state-of-the-art methods.
提出了一种有监督的方法对高光谱图像中的物质区域进行分类和分割。我们的算法是一个两阶段的过程,结合了像素分类步骤和分割步骤,旨在最小化所得区域的总周长。我们的算法在确保局部均匀区域内标签一致性和生成具有光滑边界的材料段的能力方面是独特的。此外,我们建立了一个新的高光谱基准数据集,以证明所提出的方法优于几种最先进的方法。
{"title":"Material segmentation in hyperspectral images with minimal region perimeters","authors":"Yu Zhang, C. P. Huynh, N. Habili, K. Ngan","doi":"10.1109/ICIP.2016.7532474","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532474","url":null,"abstract":"We propose a supervised approach to the classification and segmentation of material regions in hyperspectral imagery. Our algorithm is a two-stage process, combining a pixelwise classification step with a segmentation step aiming to minimise the total perimeters of the resulting regions. Our algorithm is distinctive in its ability to ensure label consistency within local homogeneous areas and to generate material segments with smooth boundaries. Furthermore, we establish a new hyperspectral benchmark dataset to demonstrate the advantages of the proposed approach over several state-of-the-art methods.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"213 1","pages":"834-838"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79768473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A shape feature based bovw method for image classification using N-gram and spatial pyramid coding scheme 基于形状特征的bovw图像分类方法,采用N-gram和空间金字塔编码方案
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532408
Elham Etemad, Gang Hu, Q. Gao
Image classification is a general visual analysis task based on the image content coded by its representation. In this research, we proposed an image representation method that is based on the perceptual shape features and their spatial distributions. A natural language processing concept, N-gram, is adopted to generate a set of perceptual shape visual words for encoding image features. By combining hierarchical visual words and spatial pyramid, Spatio-Shape Pyramid representation is constructed to reduce the semantic gaps. Experimental results show that the proposed method outperforms other state-of-the-art methods.
图像分类是一项基于图像内容编码的通用视觉分析任务。在本研究中,我们提出了一种基于感知形状特征及其空间分布的图像表示方法。采用自然语言处理概念N-gram生成一组感知形状的视觉词,用于对图像特征进行编码。将分层视觉词与空间金字塔相结合,构建空间形状金字塔表示,减少语义缺口。实验结果表明,该方法优于其他先进的方法。
{"title":"A shape feature based bovw method for image classification using N-gram and spatial pyramid coding scheme","authors":"Elham Etemad, Gang Hu, Q. Gao","doi":"10.1109/ICIP.2016.7532408","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532408","url":null,"abstract":"Image classification is a general visual analysis task based on the image content coded by its representation. In this research, we proposed an image representation method that is based on the perceptual shape features and their spatial distributions. A natural language processing concept, N-gram, is adopted to generate a set of perceptual shape visual words for encoding image features. By combining hierarchical visual words and spatial pyramid, Spatio-Shape Pyramid representation is constructed to reduce the semantic gaps. Experimental results show that the proposed method outperforms other state-of-the-art methods.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"51 1","pages":"504-508"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84938960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Measurement of critical temporal inconsistency for quality assessment of synthesized video 合成视频质量评估中关键时间不一致性的测量
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532513
Hak Gu Kim, Yong Man Ro
This paper proposes a new temporal consistency measure for quality assessment of synthesized video. Disocclusion regions appear hole regions of the synthesized video at virtual viewpoints. Filling hole regions could be problematic when the synthesized video is perceived through multi-view displays. In particular, the temporal inconsistency caused by hole filling process in view synthesis could affect the perceptual quality of the synthesized video. In the proposed method, we extract excessive flicker regions between consecutive frames and quantify the perceptual effects of the temporal inconsistency on them by measuring the structural similarity. We have demonstrated the validity of the proposed quality measure by comparisons of subjective ratings and existing objective metrics. Experimental results have shown that the proposed temporal inconsistency measure is highly correlated with the overall quality of the synthesized video.
提出了一种用于合成视频质量评价的时间一致性度量方法。在虚拟视点合成视频中,错位区域出现空洞区域。当通过多视图显示器感知合成视频时,填充孔洞区域可能会出现问题。特别是在视点合成过程中,由于补孔过程产生的时间不一致性会影响合成视频的感知质量。在该方法中,我们提取连续帧之间的过度闪烁区域,并通过测量结构相似性来量化时间不一致对它们的感知影响。我们通过比较主观评分和现有的客观指标,证明了提出的质量度量的有效性。实验结果表明,所提出的时间不一致性度量与合成视频的整体质量高度相关。
{"title":"Measurement of critical temporal inconsistency for quality assessment of synthesized video","authors":"Hak Gu Kim, Yong Man Ro","doi":"10.1109/ICIP.2016.7532513","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532513","url":null,"abstract":"This paper proposes a new temporal consistency measure for quality assessment of synthesized video. Disocclusion regions appear hole regions of the synthesized video at virtual viewpoints. Filling hole regions could be problematic when the synthesized video is perceived through multi-view displays. In particular, the temporal inconsistency caused by hole filling process in view synthesis could affect the perceptual quality of the synthesized video. In the proposed method, we extract excessive flicker regions between consecutive frames and quantify the perceptual effects of the temporal inconsistency on them by measuring the structural similarity. We have demonstrated the validity of the proposed quality measure by comparisons of subjective ratings and existing objective metrics. Experimental results have shown that the proposed temporal inconsistency measure is highly correlated with the overall quality of the synthesized video.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"41 1","pages":"1027-1031"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81767724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Weighted regularized ASM for face alignment 人脸对齐的加权正则化ASM
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532891
Guillermo Ruiz, Eduard Ramon, J. G. Giraldez, M. Ballester, F. Sukno
Active Shape Models are a powerful and well known method to perform face alignment. In some applications it is common to have shape information available beforehand, such as previously detected landmarks. Introducing this prior knowledge to the statistical model may result of great advantage but it is challenging to maintain this priors unchanged once the statistical model constraints are applied. We propose a new weighted-regularized projection into the parameter space which allows us to obtain shapes that at the same time fulfill the imposed shape constraints and are plausible according to the statistical model. The performed experiments show how using this projection better performance than competing state of the art methods is achieved.
主动形状模型是一种功能强大且众所周知的面部对齐方法。在某些应用程序中,通常预先提供形状信息,例如先前检测到的地标。将这种先验知识引入统计模型可能会带来很大的优势,但一旦应用了统计模型约束,要保持这种先验不变是一项挑战。我们提出了一种新的加权正则化投影到参数空间中,使我们能够获得同时满足所施加的形状约束并且根据统计模型是可信的形状。实验结果表明,使用这种投影方法可以获得比现有方法更好的性能。
{"title":"Weighted regularized ASM for face alignment","authors":"Guillermo Ruiz, Eduard Ramon, J. G. Giraldez, M. Ballester, F. Sukno","doi":"10.1109/ICIP.2016.7532891","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532891","url":null,"abstract":"Active Shape Models are a powerful and well known method to perform face alignment. In some applications it is common to have shape information available beforehand, such as previously detected landmarks. Introducing this prior knowledge to the statistical model may result of great advantage but it is challenging to maintain this priors unchanged once the statistical model constraints are applied. We propose a new weighted-regularized projection into the parameter space which allows us to obtain shapes that at the same time fulfill the imposed shape constraints and are plausible according to the statistical model. The performed experiments show how using this projection better performance than competing state of the art methods is achieved.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"18 1","pages":"2906-2910"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82030978","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fast multidimensional image processing with OpenCL 快速多维图像处理与OpenCL
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532664
Daniel Oliveira Dantas, H. Leal, Davy Oliveira Barros Sousa
Multidimensional image data, i.e., images with three or more dimensions, are used in many areas of science. Multidimensional image proçessing is supported in Python and MATLAB. VisionGL is an open source library that provides a set of image processing functions and can help the programmer by automatically generating code. The objective of this work is to augment VisionGL by adding multidimensional image processing support with OpenCL for high performance through use of GPUs. Benchmarking experiments were run with window and point operations to compare Python, MATLAB and VisionGL when processing 1D to 5D images. As a result, speedups of up to two orders of magnitude were obtained.
多维图像数据,即具有三维或更多维度的图像,用于许多科学领域。Python和MATLAB支持多维图像处理。VisionGL是一个开源库,它提供了一组图像处理功能,可以通过自动生成代码来帮助程序员。这项工作的目标是通过使用gpu在OpenCL中添加多维图像处理支持来增强VisionGL的高性能。通过窗口和点运算进行基准实验,比较Python、MATLAB和VisionGL在处理1D到5D图像时的性能。结果,获得了高达两个数量级的加速。
{"title":"Fast multidimensional image processing with OpenCL","authors":"Daniel Oliveira Dantas, H. Leal, Davy Oliveira Barros Sousa","doi":"10.1109/ICIP.2016.7532664","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532664","url":null,"abstract":"Multidimensional image data, i.e., images with three or more dimensions, are used in many areas of science. Multidimensional image proçessing is supported in Python and MATLAB. VisionGL is an open source library that provides a set of image processing functions and can help the programmer by automatically generating code. The objective of this work is to augment VisionGL by adding multidimensional image processing support with OpenCL for high performance through use of GPUs. Benchmarking experiments were run with window and point operations to compare Python, MATLAB and VisionGL when processing 1D to 5D images. As a result, speedups of up to two orders of magnitude were obtained.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"11 1","pages":"1779-1783"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81397215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Visual tracking with sparse correlation filters 稀疏相关滤波器的视觉跟踪
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532395
Yanmei Dong, Min Yang, Mingtao Pei
Correlation filters have recently made significant improvements in visual object tracking on both efficiency and accuracy. In this paper, we propose a sparse correlation filter, which combines the effectiveness of sparse representation and the computational efficiency of correlation filters. The sparse representation is achieved through solving an ℓ0 regularized least squares problem. The obtained sparse correlation filters are able to represent the essential information of the tracked target while being insensitive to noise. During tracking, the appearance of the target is modeled by a sparse correlation filter, and the filter is re-trained after tracking on each frame to adapt to the appearance changes of the target. The experimental results on the CVPR2013 Online Object Tracking Benchmark (OOTB) show the effectiveness of our sparse correlation filter-based tracker.
最近,相关滤波器在视觉目标跟踪的效率和准确性上都取得了显著的进步。本文提出了一种稀疏相关滤波器,它结合了稀疏表示的有效性和相关滤波器的计算效率。通过求解一个l0正则化最小二乘问题来实现稀疏表示。得到的稀疏相关滤波器既能反映被跟踪目标的基本信息,又对噪声不敏感。在跟踪过程中,利用稀疏相关滤波器对目标的外观进行建模,并在每帧跟踪后对滤波器进行重新训练,以适应目标的外观变化。在CVPR2013在线目标跟踪基准(OOTB)上的实验结果表明了我们基于稀疏相关滤波器的跟踪器的有效性。
{"title":"Visual tracking with sparse correlation filters","authors":"Yanmei Dong, Min Yang, Mingtao Pei","doi":"10.1109/ICIP.2016.7532395","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532395","url":null,"abstract":"Correlation filters have recently made significant improvements in visual object tracking on both efficiency and accuracy. In this paper, we propose a sparse correlation filter, which combines the effectiveness of sparse representation and the computational efficiency of correlation filters. The sparse representation is achieved through solving an ℓ0 regularized least squares problem. The obtained sparse correlation filters are able to represent the essential information of the tracked target while being insensitive to noise. During tracking, the appearance of the target is modeled by a sparse correlation filter, and the filter is re-trained after tracking on each frame to adapt to the appearance changes of the target. The experimental results on the CVPR2013 Online Object Tracking Benchmark (OOTB) show the effectiveness of our sparse correlation filter-based tracker.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"13 1","pages":"439-443"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82477045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Print quality assessment for stochastic clustered-dot halftones using compactness measures 用密实度评价随机聚簇网点半色调的印刷质量
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533069
P. Goyal, J. Allebach
Most electro-photographic printers prefer clustered-dot halftone textures for rendering smooth and stable prints. Clustered-dot halftone patterns can be periodic or aperiodic. As periodic clustered-dot halftone can lead to undesirable moiré patterns, stochastic clustered-dot halftone textures are more preferred. There are available different screening methods to generate stochastic clustered-dot halftone textures but there are no standard print quality assessment measures that can be easily used for quantitatively evaluating and comparing different stochastic clustered-dot halftoning methods. We explore the use of compactness measures for this purpose, and also propose a new compactness measure that seems good metric to quantitatively compare and assess the print quality of different stochastic clustered-dot halftoning methods. Using the proposed metric, we compare three different stochastic clustered-dot halftoning methods, and our results are almost in agreement with psychophysical experiments results reported earlier.
大多数电子照相打印机都喜欢簇点半色调纹理,以呈现光滑稳定的打印效果。簇点半色调图案可以是周期性的,也可以是非周期性的。由于周期性簇点半色调会导致不理想的波纹图案,因此更倾向于采用随机簇点半色调纹理。虽然产生随机簇点半色调纹理的筛选方法多种多样,但目前还没有一个标准的印刷质量评价指标,可以方便地对不同的随机簇点半色调纹理进行定量评价和比较。我们为此目的探索了紧凑度度量的使用,并提出了一种新的紧凑度度量,该度量似乎是定量比较和评估不同随机簇点半色调方法的打印质量的良好度量。利用所提出的度量,我们比较了三种不同的随机聚簇点半调方法,我们的结果与之前报道的心理物理实验结果几乎一致。
{"title":"Print quality assessment for stochastic clustered-dot halftones using compactness measures","authors":"P. Goyal, J. Allebach","doi":"10.1109/ICIP.2016.7533069","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533069","url":null,"abstract":"Most electro-photographic printers prefer clustered-dot halftone textures for rendering smooth and stable prints. Clustered-dot halftone patterns can be periodic or aperiodic. As periodic clustered-dot halftone can lead to undesirable moiré patterns, stochastic clustered-dot halftone textures are more preferred. There are available different screening methods to generate stochastic clustered-dot halftone textures but there are no standard print quality assessment measures that can be easily used for quantitatively evaluating and comparing different stochastic clustered-dot halftoning methods. We explore the use of compactness measures for this purpose, and also propose a new compactness measure that seems good metric to quantitatively compare and assess the print quality of different stochastic clustered-dot halftoning methods. Using the proposed metric, we compare three different stochastic clustered-dot halftoning methods, and our results are almost in agreement with psychophysical experiments results reported earlier.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"73 1","pages":"3792-3796"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86556339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
GCE-based model for the fusion of multiples color image segmentations 基于gce的多色图像分割融合模型
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532824
Lazhar Khelifi, M. Mignotte
In this work, we introduce a new fusion model whose objective is to fuse multiple region-based segmentation maps to get a final better segmentation result. This new fusion model is based on an energy function originated from the global consistency error (GCE), a perceptual measure which takes into account the inherent multiscale nature of an image segmentation by measuring the level of refinement existing between two spatial partitions. Combined with a region merging/splitting prior, this new energy-based fusion model of label fields allows to define an interesting penalized likelihood estimation procedure based on the global consistency error criterion with which the fusion of basic, rapidly-computed segmentation results appears as a relevant alternative compared with other segmentation techniques proposed in the image segmentation field. The performance of our fusion model was evaluated on the Berkeley dataset including various segmentations given by humans.
在本文中,我们引入了一种新的融合模型,其目标是融合多个基于区域的分割图,以获得更好的最终分割结果。这种新的融合模型基于源自全局一致性误差(GCE)的能量函数,GCE是一种感知度量,通过测量两个空间分区之间存在的细化水平来考虑图像分割固有的多尺度性质。结合区域合并/分裂先验,这种新的基于能量的标签域融合模型允许定义一个有趣的基于全局一致性错误准则的惩罚似然估计过程,与图像分割领域中提出的其他分割技术相比,融合基本的、快速计算的分割结果是一种相关的替代方案。我们的融合模型的性能在伯克利数据集上进行了评估,包括人类给出的各种分割。
{"title":"GCE-based model for the fusion of multiples color image segmentations","authors":"Lazhar Khelifi, M. Mignotte","doi":"10.1109/ICIP.2016.7532824","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532824","url":null,"abstract":"In this work, we introduce a new fusion model whose objective is to fuse multiple region-based segmentation maps to get a final better segmentation result. This new fusion model is based on an energy function originated from the global consistency error (GCE), a perceptual measure which takes into account the inherent multiscale nature of an image segmentation by measuring the level of refinement existing between two spatial partitions. Combined with a region merging/splitting prior, this new energy-based fusion model of label fields allows to define an interesting penalized likelihood estimation procedure based on the global consistency error criterion with which the fusion of basic, rapidly-computed segmentation results appears as a relevant alternative compared with other segmentation techniques proposed in the image segmentation field. The performance of our fusion model was evaluated on the Berkeley dataset including various segmentations given by humans.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"35 1","pages":"2574-2578"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86574057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
No-reference document image quality assessment based on high order image statistics 基于高阶图像统计的无参考文档图像质量评估
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532968
Jingtao Xu, Peng Ye, Qiaohong Li, Yong Liu, D. Doermann
Document image quality assessment (DIQA) aims to predict the visual quality of degraded document images. Although the definition of “visual quality” can change based on the specific applications, in this paper, we use OCR accuracy as a metric for quality and develop a novel no-reference DIQA method based on high order image statistics for OCR accuracy prediction. The proposed method consists of three steps. First, normalized local image patches are extracted with regular grid and a comprehensive document image codebook is constructed by K-means clustering. Second, local features are softly assigned to several nearest codewords, and the direct differences between high order statistics of local features and codewords are calculated as global quality aware features. Finally, support vector regression (SVR) is utilized to learn the mapping between extracted image features and OCR accuracies. Experimental results on two document image databases show that the proposed method can accurately predict OCR accuracy and outperforms previous algorithms.
文档图像质量评估(DIQA)的目的是预测退化的文档图像的视觉质量。虽然“视觉质量”的定义可能会根据具体应用而变化,但在本文中,我们将OCR精度作为质量度量,并开发了一种基于高阶图像统计的无参考DIQA方法用于OCR精度预测。该方法分为三个步骤。首先,用规则网格提取归一化的局部图像补丁,并通过K-means聚类构建完整的文档图像码本;其次,将局部特征软分配给几个最近的码字,并计算局部特征和码字的高阶统计量之间的直接差异作为全局质量感知特征;最后,利用支持向量回归(SVR)学习提取的图像特征与OCR精度之间的映射关系。在两个文档图像数据库上的实验结果表明,该方法能够准确地预测OCR精度,并优于现有算法。
{"title":"No-reference document image quality assessment based on high order image statistics","authors":"Jingtao Xu, Peng Ye, Qiaohong Li, Yong Liu, D. Doermann","doi":"10.1109/ICIP.2016.7532968","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532968","url":null,"abstract":"Document image quality assessment (DIQA) aims to predict the visual quality of degraded document images. Although the definition of “visual quality” can change based on the specific applications, in this paper, we use OCR accuracy as a metric for quality and develop a novel no-reference DIQA method based on high order image statistics for OCR accuracy prediction. The proposed method consists of three steps. First, normalized local image patches are extracted with regular grid and a comprehensive document image codebook is constructed by K-means clustering. Second, local features are softly assigned to several nearest codewords, and the direct differences between high order statistics of local features and codewords are calculated as global quality aware features. Finally, support vector regression (SVR) is utilized to learn the mapping between extracted image features and OCR accuracies. Experimental results on two document image databases show that the proposed method can accurately predict OCR accuracy and outperforms previous algorithms.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"49 1","pages":"3289-3293"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83656057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
期刊
2016 IEEE International Conference on Image Processing (ICIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1