首页 > 最新文献

2014 IEEE Visual Communications and Image Processing Conference最新文献

英文 中文
A novel metric for efficient video shot boundary detection 一种新的视频镜头边界检测方法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051500
Juan Sun, Y. Wan
With the current rate of video data generation, there is an urgent need of automatic video content analysis for subsequent purposes such as summarization, retrieval and classification. And video shot boundary detection is usually the first step to segment a video clip into meaningful shots. Taking the processing speed into account, most state-of-the-art methods make use of the frame histogram to extract shot boundary characteristics. In this paper we propose a different approach with a novel metric, which essentially captures the observation that within any shot, a pixel value in any frame usually has a pixel value very close to it within a small neighborhood in an adjacent frame. It turns out that the proposed approach can make better use of frame structural content than the histogram approach. In addition, the proposed metric has a low computational complexity. We propose a video shot boundary detection algorithm based on the proposed metric for detecting both cut transition (CT) boundary and gradual transition (GT) boundary. Experimental results show that the proposed approach enjoys better detection rates over the state-of-the-art with competitive processing speed.
以目前视频数据的生成速度,迫切需要对视频内容进行自动分析,以便进行总结、检索和分类等后续工作。视频镜头边界检测通常是将视频片段分割成有意义的镜头的第一步。考虑到处理速度,大多数最新的方法都是利用帧直方图来提取镜头边界特征。在本文中,我们提出了一种不同的方法,使用一种新的度量,它基本上捕捉到在任何镜头中,任何帧中的像素值通常在相邻帧的小邻域中都有非常接近的像素值。结果表明,该方法比直方图方法能更好地利用框架结构内容。此外,所提出的度量具有较低的计算复杂度。我们提出了一种基于所提度量的视频镜头边界检测算法,可以同时检测切移(CT)边界和渐变(GT)边界。实验结果表明,该方法具有较好的检测率和较好的处理速度。
{"title":"A novel metric for efficient video shot boundary detection","authors":"Juan Sun, Y. Wan","doi":"10.1109/VCIP.2014.7051500","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051500","url":null,"abstract":"With the current rate of video data generation, there is an urgent need of automatic video content analysis for subsequent purposes such as summarization, retrieval and classification. And video shot boundary detection is usually the first step to segment a video clip into meaningful shots. Taking the processing speed into account, most state-of-the-art methods make use of the frame histogram to extract shot boundary characteristics. In this paper we propose a different approach with a novel metric, which essentially captures the observation that within any shot, a pixel value in any frame usually has a pixel value very close to it within a small neighborhood in an adjacent frame. It turns out that the proposed approach can make better use of frame structural content than the histogram approach. In addition, the proposed metric has a low computational complexity. We propose a video shot boundary detection algorithm based on the proposed metric for detecting both cut transition (CT) boundary and gradual transition (GT) boundary. Experimental results show that the proposed approach enjoys better detection rates over the state-of-the-art with competitive processing speed.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124622840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Implementation of fast free-viewpoint video rendering on graphics processing units 在图形处理单元上实现快速自由视点视频渲染
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051585
Nicholas Attard, C. J. Debono
This paper presents the development of a fast Free-Viewpoint Video (FW) rendering algorithm that exploits the parallelism offered by General Purpose Graphics Processing Units (GPGPUs). The system generates virtual views through the use of Depth Image-Based Rendering (DIBR) algorithms, implemented using NVidia® Compute Unified Device Architecture (CUDA). A novel reference image brightness adjustment algorithm that exploits the correspondences between matching pixels in the reference images to avoid drastic brightness switching while navigating in between views is also discussed. The developed solution ensures that data transfers are kept at a minimum, thus improving the overall rendering speed. Objective and subjective test results show that, for typical free-view scenarios, the proposed algorithm can be successfully deployed in real-time FW systems, providing a good Quality of Experience (QoE).
本文提出了一种利用通用图形处理单元(gpgpu)提供的并行性开发的快速自由视点视频(FW)渲染算法。该系统通过使用深度图像渲染(DIBR)算法生成虚拟视图,该算法使用NVidia®计算统一设备架构(CUDA)实现。本文还讨论了一种新的参考图像亮度调整算法,该算法利用参考图像中匹配像素之间的对应关系来避免在视图之间导航时剧烈的亮度切换。开发的解决方案确保数据传输保持在最低限度,从而提高整体呈现速度。客观和主观测试结果表明,在典型的自由视角场景下,该算法可以成功地部署在实时FW系统中,提供了良好的体验质量(QoE)。
{"title":"Implementation of fast free-viewpoint video rendering on graphics processing units","authors":"Nicholas Attard, C. J. Debono","doi":"10.1109/VCIP.2014.7051585","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051585","url":null,"abstract":"This paper presents the development of a fast Free-Viewpoint Video (FW) rendering algorithm that exploits the parallelism offered by General Purpose Graphics Processing Units (GPGPUs). The system generates virtual views through the use of Depth Image-Based Rendering (DIBR) algorithms, implemented using NVidia® Compute Unified Device Architecture (CUDA). A novel reference image brightness adjustment algorithm that exploits the correspondences between matching pixels in the reference images to avoid drastic brightness switching while navigating in between views is also discussed. The developed solution ensures that data transfers are kept at a minimum, thus improving the overall rendering speed. Objective and subjective test results show that, for typical free-view scenarios, the proposed algorithm can be successfully deployed in real-time FW systems, providing a good Quality of Experience (QoE).","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124695541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
1-D dictionary mode for screen content coding 用于屏幕内容编码的一维字典模式
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051536
Bin Li, Jizheng Xu, Feng Wu
This paper introduces 1-D dictionary mode designed for screen content coding. Two 1-D dictionary modes are designed to improve the coding efficiency for screen content. The first one is called normal dictionary mode, in which a virtual dictionary should be maintained and all the prediction comes from the virtual dictionary. The other one is called reconstruction based dictionary mode, where no virtual dictionary is to be maintained and all the previously reconstructed pixels in the same picture can be used for prediction. Hash based search is designed to find matching for both dictionary modes efficiently. 1-D dictionary mode with variable block sizes are also supported in the proposed scheme. The experimental results show the proposed algorithm achieves about 10% ~ 18.4% bit saving for different coding structures. The bit saving is up to 60% for the proposed method.
介绍了一种用于屏幕内容编码的一维字典模式。为了提高屏幕内容的编码效率,设计了两种一维字典模式。第一种是正常字典模式,即维护一个虚拟字典,所有的预测都来自虚拟字典。另一种是基于重建的字典模式,这种模式不需要维护虚拟字典,可以使用同一张图片中所有先前重建的像素进行预测。基于哈希的搜索旨在有效地找到两种字典模式的匹配。该方案还支持可变块大小的一维字典模式。实验结果表明,对于不同的编码结构,该算法可以节省10% ~ 18.4%的比特。该方法可节省高达60%的钻头。
{"title":"1-D dictionary mode for screen content coding","authors":"Bin Li, Jizheng Xu, Feng Wu","doi":"10.1109/VCIP.2014.7051536","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051536","url":null,"abstract":"This paper introduces 1-D dictionary mode designed for screen content coding. Two 1-D dictionary modes are designed to improve the coding efficiency for screen content. The first one is called normal dictionary mode, in which a virtual dictionary should be maintained and all the prediction comes from the virtual dictionary. The other one is called reconstruction based dictionary mode, where no virtual dictionary is to be maintained and all the previously reconstructed pixels in the same picture can be used for prediction. Hash based search is designed to find matching for both dictionary modes efficiently. 1-D dictionary mode with variable block sizes are also supported in the proposed scheme. The experimental results show the proposed algorithm achieves about 10% ~ 18.4% bit saving for different coding structures. The bit saving is up to 60% for the proposed method.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"103 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124754849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Studying the added value of computational saliency in objective image quality assessment 研究了计算显著性在客观图像质量评价中的附加价值
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051494
Wei Zhang, A. Borji, Fuzheng Yang, Ping Jiang, Hantao Liu
Advances in image quality assessment have shown the potential added value of including visual attention aspects in objective quality metrics. Numerous models of visual saliency are implemented and integrated in different quality metrics; however, their ability of improving a metric's performance in predicting perceived image quality is not fully investigated. In this paper, we conduct an exhaustive comparison of 20 state-of-the-art saliency models in the context of image quality assessment. Experimental results show that adding computational saliency is beneficial to quality prediction in general terms. However, the amount of performance gain that can be obtained by adding saliency in quality metrics highly depends on the saliency model and on the metric.
图像质量评估的进展表明,在客观质量度量中包括视觉注意方面的潜在附加价值。在不同的质量度量中实现和集成了许多视觉显著性模型;然而,它们在预测感知图像质量方面提高度量性能的能力并没有得到充分的研究。在本文中,我们在图像质量评估的背景下,对20个最先进的显著性模型进行了详尽的比较。实验结果表明,一般来说,增加计算显著性有利于质量预测。然而,通过在质量度量中添加显著性而获得的性能增益在很大程度上取决于显著性模型和度量。
{"title":"Studying the added value of computational saliency in objective image quality assessment","authors":"Wei Zhang, A. Borji, Fuzheng Yang, Ping Jiang, Hantao Liu","doi":"10.1109/VCIP.2014.7051494","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051494","url":null,"abstract":"Advances in image quality assessment have shown the potential added value of including visual attention aspects in objective quality metrics. Numerous models of visual saliency are implemented and integrated in different quality metrics; however, their ability of improving a metric's performance in predicting perceived image quality is not fully investigated. In this paper, we conduct an exhaustive comparison of 20 state-of-the-art saliency models in the context of image quality assessment. Experimental results show that adding computational saliency is beneficial to quality prediction in general terms. However, the amount of performance gain that can be obtained by adding saliency in quality metrics highly depends on the saliency model and on the metric.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"259 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123967130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Comparative study of 8 and 10-bit HEVC encoders 8位和10位HEVC编码器的比较研究
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051626
Jarno Vanne, Marko Viitanen, Ari Koivula, T. Hämäläinen
This paper compares the rate-distortion-complexity (RDC) characteristics of the HEVC Main 10 Profile (M10P) and Main Profile (MP) encoders. The evaluations are performed with HEVC reference encoder (HM) whose M10P and MP are benchmarked with different resolutions, frame rates, and bit depths. The reported RD results are based on bit rate differences for equal PSNR whereas complexities have been profiled with Intel VTune on Intel Core 2 processor. With our 10-bit 4K 120 fps test set, the average bit rate decrements of M10P over MP are 5.8%, 11.6%, and 12.3% in the all-intra (AI), random access (RA), and low-delay B (LB) configurations, respectively. Decreasing the bit depth of this test set to 8 lowers the RD gain of Ml OP only slightly to 5.4% (AI), 11.4% (RA), and 12.1% (LB). The similar trend continues in all our tests even though the RD gain of M10P is decreased over MP with lower resolutions and frame rates. M10P introduces no computational overhead in HM, but it is anticipated to increase complexity and double the memory usage in practical encoders. Hence, the 10-bit HEVC encoding with 8-bit input video is the most recommended option if computation and memory resources are adequate for it.
本文比较了HEVC主配置文件(M10P)和主配置文件(MP)编码器的码率失真复杂度(RDC)特性。使用HEVC参考编码器(HM)进行评估,其M10P和MP以不同的分辨率,帧速率和位深度为基准。报告的RD结果是基于相同PSNR的比特率差异,而复杂性是在英特尔酷睿2处理器上使用英特尔VTune进行的。使用我们的10位4K 120 fps测试集,在全内(AI)、随机访问(RA)和低延迟B (LB)配置下,M10P比MP的平均比特率下降分别为5.8%、11.6%和12.3%。将该测试集的位深降低为8,Ml OP的RD增益仅略微降低至5.4% (AI), 11.4% (RA)和12.1% (LB)。在我们所有的测试中,尽管M10P的RD增益比分辨率和帧率较低的MP低,但类似的趋势仍在继续。M10P在HM中没有引入计算开销,但预计会增加复杂性,并使实际编码器的内存使用量增加一倍。因此,如果计算和内存资源足够的话,10位HEVC编码和8位输入视频是最推荐的选择。
{"title":"Comparative study of 8 and 10-bit HEVC encoders","authors":"Jarno Vanne, Marko Viitanen, Ari Koivula, T. Hämäläinen","doi":"10.1109/VCIP.2014.7051626","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051626","url":null,"abstract":"This paper compares the rate-distortion-complexity (RDC) characteristics of the HEVC Main 10 Profile (M10P) and Main Profile (MP) encoders. The evaluations are performed with HEVC reference encoder (HM) whose M10P and MP are benchmarked with different resolutions, frame rates, and bit depths. The reported RD results are based on bit rate differences for equal PSNR whereas complexities have been profiled with Intel VTune on Intel Core 2 processor. With our 10-bit 4K 120 fps test set, the average bit rate decrements of M10P over MP are 5.8%, 11.6%, and 12.3% in the all-intra (AI), random access (RA), and low-delay B (LB) configurations, respectively. Decreasing the bit depth of this test set to 8 lowers the RD gain of Ml OP only slightly to 5.4% (AI), 11.4% (RA), and 12.1% (LB). The similar trend continues in all our tests even though the RD gain of M10P is decreased over MP with lower resolutions and frame rates. M10P introduces no computational overhead in HM, but it is anticipated to increase complexity and double the memory usage in practical encoders. Hence, the 10-bit HEVC encoding with 8-bit input video is the most recommended option if computation and memory resources are adequate for it.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114712758","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Patch-based image deblocking using geodesic distance weighted low-rank approximation 基于测地线距离加权低秩近似的图像块化方法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051514
Mading Li, Jiaying Liu, J. Ren, Zongming Guo
Transform coding based on the discrete cosine transform (DCT) has been widely used in image coding standards. However, the coded images often suffer from severe visual distortions such as blocking artifacts. In this paper, we propose a novel image deblocking method to address the blocking artifacts reduction problem in a patch-based scheme. Image patches are clustered and reconstructed by the low-rank approximation, which is weighted by the geodesic distance. Experimental results show that the proposed method achieves higher PSNR than the state-of-the-art deblocking and denoising methods and the processed images present good visual quality.
基于离散余弦变换(DCT)的变换编码在图像编码标准中得到了广泛应用。然而,编码后的图像往往会出现严重的视觉失真,如阻塞伪影。在本文中,我们提出了一种新的图像块化方法来解决基于补丁的方案中块伪影的减少问题。通过低秩近似对图像块进行聚类和重构,低秩近似由测地线距离加权。实验结果表明,该方法比现有的去块和去噪方法具有更高的PSNR,处理后的图像具有良好的视觉质量。
{"title":"Patch-based image deblocking using geodesic distance weighted low-rank approximation","authors":"Mading Li, Jiaying Liu, J. Ren, Zongming Guo","doi":"10.1109/VCIP.2014.7051514","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051514","url":null,"abstract":"Transform coding based on the discrete cosine transform (DCT) has been widely used in image coding standards. However, the coded images often suffer from severe visual distortions such as blocking artifacts. In this paper, we propose a novel image deblocking method to address the blocking artifacts reduction problem in a patch-based scheme. Image patches are clustered and reconstructed by the low-rank approximation, which is weighted by the geodesic distance. Experimental results show that the proposed method achieves higher PSNR than the state-of-the-art deblocking and denoising methods and the processed images present good visual quality.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114823166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Dynamic redundancy allocation for video streaming using Sub-GOP based FEC code 基于Sub-GOP的FEC代码的视频流动态冗余分配
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051620
Li Yu, Jimin Xiao, T. Tillo
Reed-Solomon erasure code is one of the most studied protection methods for video streaming over unreliable networks. As a block-based error correcting code, large block size and increased number of parity packets will enhance its protection performance. However, for video applications this enhancement is sacrificed by the error propagation and the increased bitrate. So, to tackle this paradox, we propose a rate-distortion optimized redundancy allocation scheme, which takes into consideration the distortion caused by losing each slice and the propagated error. Different from other approaches, the amount of introduced redundancy and the way it is introduced are automatically selected without human interventions based on the network condition and video characteristics. The redundancy allocation problem is formulated as a constraint optimization problem, which allows to have more flexibility in setting the block-wise redundancy. The proposed scheme is implemented in JM14.0 for H.264, and it achieves an average gain of 1dB over the state-of-the-art approach.
里德-所罗门擦除码是研究最多的不可靠网络视频流保护方法之一。作为一种基于块的纠错码,块大小越大,奇偶校验包数量越多,其保护性能越好。然而,对于视频应用,这种增强被错误传播和增加的比特率所牺牲。因此,为了解决这一矛盾,我们提出了一种率失真优化的冗余分配方案,该方案考虑了丢失每个片引起的失真和传播误差。与其他方法不同的是,引入冗余的数量和引入方式是根据网络条件和视频特征自动选择的,无需人工干预。将冗余分配问题表述为约束优化问题,使得在设置块级冗余时具有更大的灵活性。该方案在H.264的JM14.0中实现,与最先进的方法相比,平均增益为1dB。
{"title":"Dynamic redundancy allocation for video streaming using Sub-GOP based FEC code","authors":"Li Yu, Jimin Xiao, T. Tillo","doi":"10.1109/VCIP.2014.7051620","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051620","url":null,"abstract":"Reed-Solomon erasure code is one of the most studied protection methods for video streaming over unreliable networks. As a block-based error correcting code, large block size and increased number of parity packets will enhance its protection performance. However, for video applications this enhancement is sacrificed by the error propagation and the increased bitrate. So, to tackle this paradox, we propose a rate-distortion optimized redundancy allocation scheme, which takes into consideration the distortion caused by losing each slice and the propagated error. Different from other approaches, the amount of introduced redundancy and the way it is introduced are automatically selected without human interventions based on the network condition and video characteristics. The redundancy allocation problem is formulated as a constraint optimization problem, which allows to have more flexibility in setting the block-wise redundancy. The proposed scheme is implemented in JM14.0 for H.264, and it achieves an average gain of 1dB over the state-of-the-art approach.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124347074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A new framework for image impulse noise removal with postprocessing 基于后处理的图像脉冲噪声去除新框架
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051601
Qiqiang Chen, Y. Wan
Impulse noise is commonly encountered during image transmission and many methods have been proposed to remove it. Although it is now possible to recover the true image reasonably well, even under severe noise (90% pixel contamination), essentially all methods published so far follow the standard procedure of noisy pixel detection/classification and then noisy pixel value reconstruction, without any further processing. In this paper we show an interesting empirical discovery that the traditionally denoised image tends to have the estimation error with a Laplacian distribution, which makes it possible to add a postprocessing stage to denoise the traditionally obtained result with this new type of noise. We propose a practical algorithm within this new framework and experimental results show that superior results can be obtained over previously published methods.
脉冲噪声是图像传输过程中经常遇到的问题,人们提出了许多消除脉冲噪声的方法。虽然现在有可能很好地恢复真实图像,即使在严重的噪声(90%的像素污染)下,但迄今为止发表的所有方法基本上都遵循了噪声像素检测/分类然后噪声像素值重建的标准程序,没有任何进一步的处理。在本文中,我们展示了一个有趣的经验发现,即传统降噪的图像往往具有拉普拉斯分布的估计误差,这使得可以添加一个后处理阶段来用这种新型噪声对传统获得的结果进行降噪。在此框架下,我们提出了一种实用的算法,实验结果表明,与先前发表的方法相比,可以获得更好的结果。
{"title":"A new framework for image impulse noise removal with postprocessing","authors":"Qiqiang Chen, Y. Wan","doi":"10.1109/VCIP.2014.7051601","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051601","url":null,"abstract":"Impulse noise is commonly encountered during image transmission and many methods have been proposed to remove it. Although it is now possible to recover the true image reasonably well, even under severe noise (90% pixel contamination), essentially all methods published so far follow the standard procedure of noisy pixel detection/classification and then noisy pixel value reconstruction, without any further processing. In this paper we show an interesting empirical discovery that the traditionally denoised image tends to have the estimation error with a Laplacian distribution, which makes it possible to add a postprocessing stage to denoise the traditionally obtained result with this new type of noise. We propose a practical algorithm within this new framework and experimental results show that superior results can be obtained over previously published methods.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124385764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Edge-aware depth completion for point-cloud 3D scene visualization on an RGB-D camera RGB-D相机上点云3D场景可视化的边缘感知深度补全
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051596
Yung-Lin Huang, Tang-Wei Hsu, Shao-Yi Chien
Nowadays, 3D scene reconstruction using RGB-D videos becomes more popular because of the widely-available off-the-shelf RGB-D camera. However, the depth information from current RGB-D camera still need improved in order to reconstruct the 3D scene with better quality. In this paper, an edge-aware depth completion method aims to recover more accurate depth information is proposed. There are mainly two parts in our proposed method. The first part is the edge-aware color image analysis, and the second part is depth image processing including unreliable depth pixel invalidation and filling. The depth image processing can retrieve more accurate depth information using our proposed edge-aware color image analysis. Consequently, we can not only preserve the reliable depth information, but also fill in the appropriate depth values to align edges of depth image with edges of its corresponding color image. Besides, the experimental results show that the visualization of the reconstructed point-cloud 3D scene benefits from our proposed edge-aware depth completion. Finally, the PSNR evaluation using ground truth depth information is presented.
如今,使用RGB-D视频进行3D场景重建变得更加流行,因为现成的RGB-D相机广泛可用。但是,当前RGB-D相机的深度信息还需要改进,才能更好地重建出高质量的3D场景。本文提出了一种边缘感知深度补全方法,以获得更精确的深度信息。我们提出的方法主要分为两部分。第一部分是边缘感知彩色图像分析,第二部分是深度图像处理,包括不可靠深度像素失效和填充。利用我们提出的边缘感知彩色图像分析,深度图像处理可以获得更准确的深度信息。这样不仅可以保留可靠的深度信息,还可以填充适当的深度值,使深度图像的边缘与相应的彩色图像的边缘对齐。此外,实验结果表明,我们提出的边缘感知深度补全方法有利于重建点云三维场景的可视化。最后,提出了利用地真深度信息评价PSNR的方法。
{"title":"Edge-aware depth completion for point-cloud 3D scene visualization on an RGB-D camera","authors":"Yung-Lin Huang, Tang-Wei Hsu, Shao-Yi Chien","doi":"10.1109/VCIP.2014.7051596","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051596","url":null,"abstract":"Nowadays, 3D scene reconstruction using RGB-D videos becomes more popular because of the widely-available off-the-shelf RGB-D camera. However, the depth information from current RGB-D camera still need improved in order to reconstruct the 3D scene with better quality. In this paper, an edge-aware depth completion method aims to recover more accurate depth information is proposed. There are mainly two parts in our proposed method. The first part is the edge-aware color image analysis, and the second part is depth image processing including unreliable depth pixel invalidation and filling. The depth image processing can retrieve more accurate depth information using our proposed edge-aware color image analysis. Consequently, we can not only preserve the reliable depth information, but also fill in the appropriate depth values to align edges of depth image with edges of its corresponding color image. Besides, the experimental results show that the visualization of the reconstructed point-cloud 3D scene benefits from our proposed edge-aware depth completion. Finally, the PSNR evaluation using ground truth depth information is presented.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"452 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125786918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Growcut-based drusen segmentation for age-related macular degeneration detection 基于生长切块的黄斑变性检测的病灶分割
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051529
Huiying Liu, Yanwu Xu, D. Wong, Jiang Liu
Age-related Macular Degeneration (AMD) is the third leading cause of blindness. Its prevalence is increasing in these years for the coming of "aging time". Early detection and grading can prohibit it from becoming severe and protect vision. The appearance of drusen is an important indicator for AMD thus automatic drusen detection and segmentation have attracted much research attention in the past years. In this paper, we propose a novel drusen segmentation method by using Growcut. This method first detects the local maximum and minimum points. The maximum points, which are potential drusen, are then classified as drusen or non-drusen. The drusen points will be used as foreground labels while the non-drusen points together with the minima will be used as background labels. These labels are fed into Growcut to obtain the drusen boundaries. The method is tested on a manually labeled dataset with 96 images containing drusen. The experimental results verify the effectiveness of the method.
老年性黄斑变性(AMD)是致盲的第三大原因。近年来,随着“老龄化时代”的到来,它的患病率正在上升。早期发现和分级可以防止病情恶化,保护视力。色斑的出现是AMD的一个重要指标,因此色斑的自动检测与分割成为近年来研究的热点。本文提出了一种新的基于Growcut的样本分割方法。该方法首先检测局部最大值和最小值点。最大的点是潜在的毒品,然后被分为毒品和非毒品。drusen点将被用作前景标签,而非drusen点和最小值将被用作背景标签。这些标签被输入到Growcut中以获得酒的边界。该方法在一个手工标记的数据集上进行了测试,该数据集包含96张包含drusen的图像。实验结果验证了该方法的有效性。
{"title":"Growcut-based drusen segmentation for age-related macular degeneration detection","authors":"Huiying Liu, Yanwu Xu, D. Wong, Jiang Liu","doi":"10.1109/VCIP.2014.7051529","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051529","url":null,"abstract":"Age-related Macular Degeneration (AMD) is the third leading cause of blindness. Its prevalence is increasing in these years for the coming of \"aging time\". Early detection and grading can prohibit it from becoming severe and protect vision. The appearance of drusen is an important indicator for AMD thus automatic drusen detection and segmentation have attracted much research attention in the past years. In this paper, we propose a novel drusen segmentation method by using Growcut. This method first detects the local maximum and minimum points. The maximum points, which are potential drusen, are then classified as drusen or non-drusen. The drusen points will be used as foreground labels while the non-drusen points together with the minima will be used as background labels. These labels are fed into Growcut to obtain the drusen boundaries. The method is tested on a manually labeled dataset with 96 images containing drusen. The experimental results verify the effectiveness of the method.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"152 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122050647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2014 IEEE Visual Communications and Image Processing Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1