首页 > 最新文献

2013 Visual Communications and Image Processing (VCIP)最新文献

英文 中文
Shaking video synthesis for video stabilization performance assessment 用于视频防抖性能评价的抖动视频合成
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706422
Hui Qu, Li Song, Gengjian Xue
The goal of video stabilization is to remove the unwanted camera motion and obtain stable versions. Theoretically, a good stabilization algorithm should remove the unwanted motion without the loss of image qualities. However, due to the lack of ground-truth video frames, the accurate performance evaluation of different algorithms is hard. Most existing evaluation techniques usually synthesize stable videos from shaking ones, but they are not effective enough. Different from previous methods, in this paper we propose a novel method which synthesize shaking videos from stable frames. Based on the synthetic shaking videos, we perform preliminary video stabilization performance assessment on three stabilization algorithms. Our shaking video synthesis method can not only give a benchmark for full-reference video stabilization performance assessment, but also provide a basis for exploring the theoretical bound of video stabilization which may help to improve existing stabilization algorithms.
视频防抖的目的是消除不必要的摄像机运动,获得稳定的版本。从理论上讲,一个好的稳定算法应该在不损失图像质量的情况下消除不必要的运动。然而,由于缺乏真实视频帧,很难对不同算法进行准确的性能评估。现有的评价技术大多是由震动视频合成稳定视频,但效果不理想。与以往的方法不同,本文提出了一种从稳定帧合成抖动视频的新方法。在合成震动视频的基础上,对三种稳定算法进行了初步的视频稳定性能评价。本文提出的抖动视频合成方法不仅可以为全参考视频防抖性能评估提供基准,而且可以为探索视频防抖的理论边界提供依据,有助于改进现有的防抖算法。
{"title":"Shaking video synthesis for video stabilization performance assessment","authors":"Hui Qu, Li Song, Gengjian Xue","doi":"10.1109/VCIP.2013.6706422","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706422","url":null,"abstract":"The goal of video stabilization is to remove the unwanted camera motion and obtain stable versions. Theoretically, a good stabilization algorithm should remove the unwanted motion without the loss of image qualities. However, due to the lack of ground-truth video frames, the accurate performance evaluation of different algorithms is hard. Most existing evaluation techniques usually synthesize stable videos from shaking ones, but they are not effective enough. Different from previous methods, in this paper we propose a novel method which synthesize shaking videos from stable frames. Based on the synthetic shaking videos, we perform preliminary video stabilization performance assessment on three stabilization algorithms. Our shaking video synthesis method can not only give a benchmark for full-reference video stabilization performance assessment, but also provide a basis for exploring the theoretical bound of video stabilization which may help to improve existing stabilization algorithms.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122225887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
A parallel root-finding method for omnidirectional image unwrapping 面向全向图像展开的并行寻根方法
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706340
N. Chong, M. D. Wong, Y. Kho
The panoramic unwrapping of catadioptric omnidirectional view (COV) sensors have mostly relied on a precomputed mapping look-up table due to an expensive computational load that generally has its bottleneck occur at solving a sextic polynomial. However, this approach causes a limitation to the viewpoint dynamics as runtime modifications to the mapping values are not allowed in the implementation. In this paper, a parallel root-finding technique using Compute Unified Device Architecture (CUDA) platform is proposed. The proposed method enables on-the-fly computation of the mapping look-up table thus facilitate in a real-time viewpoint adjustable panoramic unwrapping. Experimental results showed that the proposed implementation incurred minimum computational load, and performed at 10.3 times and 2.3 times the speed of a current generation central processing unit (CPU) respectively on a single-core and multi-core environment.
反射全向视图(COV)传感器的全景展开主要依赖于预先计算的映射查找表,由于计算量大,其瓶颈通常出现在求解六次多项式时。然而,这种方法会对视点动态造成限制,因为在实现中不允许对映射值进行运行时修改。本文提出了一种基于计算统一设备架构(CUDA)平台的并行寻根技术。该方法实现了地图查找表的实时计算,实现了视点可调全景展开。实验结果表明,在单核和多核环境下,该实现的计算负荷最小,执行速度分别是当前一代中央处理器(CPU)的10.3倍和2.3倍。
{"title":"A parallel root-finding method for omnidirectional image unwrapping","authors":"N. Chong, M. D. Wong, Y. Kho","doi":"10.1109/VCIP.2013.6706340","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706340","url":null,"abstract":"The panoramic unwrapping of catadioptric omnidirectional view (COV) sensors have mostly relied on a precomputed mapping look-up table due to an expensive computational load that generally has its bottleneck occur at solving a sextic polynomial. However, this approach causes a limitation to the viewpoint dynamics as runtime modifications to the mapping values are not allowed in the implementation. In this paper, a parallel root-finding technique using Compute Unified Device Architecture (CUDA) platform is proposed. The proposed method enables on-the-fly computation of the mapping look-up table thus facilitate in a real-time viewpoint adjustable panoramic unwrapping. Experimental results showed that the proposed implementation incurred minimum computational load, and performed at 10.3 times and 2.3 times the speed of a current generation central processing unit (CPU) respectively on a single-core and multi-core environment.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115848525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Salient object detection in image sequences via spatial-temporal cue 基于时空线索的图像序列显著目标检测
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706438
Chuang Gan, Zengchang Qin, Jia Xu, T. Wan
Contemporary video search and categorization are non-trivial tasks due to the massively increasing amount and content variety of videos. We put forward the study of visual saliency models in video. Such a model is employed to identify salient objects from the image background. Starting from the observation that motion information in video often attracts more human attention compared to static images, we devise a region contrast based saliency detection model using spatial-temporal cues (RCST). We introduce and study four saliency principles to realize the RCST. This generalizes the previous static image for saliency computational model to video. We conduct experiments on a publicly available video segmentation database where our method significantly outperforms seven state-of-the-art methods with respect to PR curve, ROC curve and visual comparison.
由于视频数量和内容的大量增加,当代视频搜索和分类是一项非常重要的任务。提出了视频中视觉显著性模型的研究。该模型用于从图像背景中识别显著目标。从观察到视频中的运动信息往往比静态图像更能吸引人们的注意力开始,我们设计了一个基于区域对比度的显著性检测模型,该模型使用时空线索(RCST)。我们介绍并研究了实现RCST的四种显著性原则。将以往的静态图像显著性计算模型推广到视频中。我们在一个公开可用的视频分割数据库上进行实验,我们的方法在PR曲线、ROC曲线和视觉比较方面明显优于7种最先进的方法。
{"title":"Salient object detection in image sequences via spatial-temporal cue","authors":"Chuang Gan, Zengchang Qin, Jia Xu, T. Wan","doi":"10.1109/VCIP.2013.6706438","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706438","url":null,"abstract":"Contemporary video search and categorization are non-trivial tasks due to the massively increasing amount and content variety of videos. We put forward the study of visual saliency models in video. Such a model is employed to identify salient objects from the image background. Starting from the observation that motion information in video often attracts more human attention compared to static images, we devise a region contrast based saliency detection model using spatial-temporal cues (RCST). We introduce and study four saliency principles to realize the RCST. This generalizes the previous static image for saliency computational model to video. We conduct experiments on a publicly available video segmentation database where our method significantly outperforms seven state-of-the-art methods with respect to PR curve, ROC curve and visual comparison.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121120919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-scale video text detection based on corner and stroke width verification 基于角和笔画宽度验证的多尺度视频文本检测
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706387
Boyu Zhang, Jiafeng Liu, Xianglong Tang
Focusing on the video text detection, which is challenging and with wide potential applications, a novel stroke width feature is proposed and a system which detects text regions based on multi-scale corner detection is implemented in this paper. In our system, candidate text regions are generated by applying morphologic operation based on corner points detected in different scales, and non-text regions are filtered by combining proposed stroke width feature with some simple geometric properties. Moreover, there is a new multi-instance semi-supervised learning strategy being proposed in this paper considering the unknown contrast parameter in stroke width extraction. Experiments taken on video frames from different kinds of video shots prove that the proposed approach is both efficient and accurate for video text detection.
针对视频文本检测这一具有挑战性和广阔应用前景的问题,提出了一种新的笔画宽度特征,并实现了一种基于多尺度角点检测的文本区域检测系统。在我们的系统中,基于不同尺度检测到的角点进行形态学运算生成候选文本区域,并结合提出的笔画宽度特征和一些简单的几何属性来过滤非文本区域。此外,本文还提出了一种考虑笔画宽度提取中对比度参数未知的多实例半监督学习策略。对不同类型的视频帧进行了实验,证明了该方法对视频文本检测的有效性和准确性。
{"title":"Multi-scale video text detection based on corner and stroke width verification","authors":"Boyu Zhang, Jiafeng Liu, Xianglong Tang","doi":"10.1109/VCIP.2013.6706387","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706387","url":null,"abstract":"Focusing on the video text detection, which is challenging and with wide potential applications, a novel stroke width feature is proposed and a system which detects text regions based on multi-scale corner detection is implemented in this paper. In our system, candidate text regions are generated by applying morphologic operation based on corner points detected in different scales, and non-text regions are filtered by combining proposed stroke width feature with some simple geometric properties. Moreover, there is a new multi-instance semi-supervised learning strategy being proposed in this paper considering the unknown contrast parameter in stroke width extraction. Experiments taken on video frames from different kinds of video shots prove that the proposed approach is both efficient and accurate for video text detection.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121827915","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Mid-level feature based local descriptor selection for image search 图像搜索中基于中级特征的局部描述符选择
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706455
S. Bucak, A. Saxena, Abhishek Nagar, Felix C. A. Fernandes, Kong-Posh Bhat
The objective in developing compact descriptors for visual image search is building an image retrieval system that works efficiently and effectively under bandwidth and memory constraints. Selecting local descriptors to be processed, and sending them to the server for matching is an integral part of such a system. One such image search and retrieval system is the Compact Descriptors for Visual Search (CDVS) standardization test model being developed by MPEG which has an efficient local descriptor selection criteria. However, all the existing selection parameters in CDVS are based on low-level features. In this paper, we propose two “mid-level” local descriptor selection criteria: Visual Meaning Score (VMS), and Visual Vocabulary Score (VVS) which can be seamlessly integrated into the existing CDVS framework. A mid-level criteria explicitly allows selection of local descriptors closer to a given set of images. Both VMS and VVS are based on visual words (patches) of images, and provide significant gains over the current CDVS standard in terms of matching accuracy, and have very low implementation cost.
开发用于视觉图像搜索的紧凑描述符的目标是建立一个在带宽和内存限制下高效工作的图像检索系统。选择要处理的局部描述符,并将它们发送到服务器进行匹配,是这样一个系统的组成部分。其中一个图像搜索和检索系统是由MPEG开发的压缩视觉搜索描述符(CDVS)标准化测试模型,该模型具有高效的局部描述符选择标准。然而,现有的cddvs选择参数都是基于底层特征的。在本文中,我们提出了两个“中级”局部描述符选择标准:视觉意义评分(VMS)和视觉词汇评分(VVS),这两个标准可以无缝集成到现有的cddvs框架中。中级标准明确地允许选择更接近给定图像集的局部描述符。VMS和VVS都是基于图像的视觉词(补丁),在匹配精度方面比目前的CDVS标准有显著提高,并且实现成本非常低。
{"title":"Mid-level feature based local descriptor selection for image search","authors":"S. Bucak, A. Saxena, Abhishek Nagar, Felix C. A. Fernandes, Kong-Posh Bhat","doi":"10.1109/VCIP.2013.6706455","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706455","url":null,"abstract":"The objective in developing compact descriptors for visual image search is building an image retrieval system that works efficiently and effectively under bandwidth and memory constraints. Selecting local descriptors to be processed, and sending them to the server for matching is an integral part of such a system. One such image search and retrieval system is the Compact Descriptors for Visual Search (CDVS) standardization test model being developed by MPEG which has an efficient local descriptor selection criteria. However, all the existing selection parameters in CDVS are based on low-level features. In this paper, we propose two “mid-level” local descriptor selection criteria: Visual Meaning Score (VMS), and Visual Vocabulary Score (VVS) which can be seamlessly integrated into the existing CDVS framework. A mid-level criteria explicitly allows selection of local descriptors closer to a given set of images. Both VMS and VVS are based on visual words (patches) of images, and provide significant gains over the current CDVS standard in terms of matching accuracy, and have very low implementation cost.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114875313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Long-term background memory based on Gaussian mixture model 基于高斯混合模型的长期背景记忆
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706397
W. Zhao, X. D. Zhao, W. M. Liu, X. L. Tang
This paper aims to present a long-term background memory framework, which is capable of memorizing long period background in video and rapidly adapting to the changes of background. Based on Gaussian mixture model (GMM), this framework enables an accurate identification of long period background appearances and presents a perfect solution to numerous typical problems on foreground detection. The experimental results with various benchmark sequences quantitatively and qualitatively demonstrate that the proposed algorithm outperforms many GMM-based methods for foreground detection, as well as other representative approaches.
本文旨在提出一种长时间背景记忆框架,能够记忆视频中的长时间背景,并能快速适应背景的变化。该框架基于高斯混合模型(GMM),能够准确地识别长周期背景,为前景检测中的许多典型问题提供了完美的解决方案。各种基准序列的定量和定性实验结果表明,该算法在前景检测方面优于许多基于gmm的方法,以及其他代表性方法。
{"title":"Long-term background memory based on Gaussian mixture model","authors":"W. Zhao, X. D. Zhao, W. M. Liu, X. L. Tang","doi":"10.1109/VCIP.2013.6706397","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706397","url":null,"abstract":"This paper aims to present a long-term background memory framework, which is capable of memorizing long period background in video and rapidly adapting to the changes of background. Based on Gaussian mixture model (GMM), this framework enables an accurate identification of long period background appearances and presents a perfect solution to numerous typical problems on foreground detection. The experimental results with various benchmark sequences quantitatively and qualitatively demonstrate that the proposed algorithm outperforms many GMM-based methods for foreground detection, as well as other representative approaches.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117186549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A local shape descriptor for mobile linedrawing retrieval 用于移动线条检索的局部形状描述符
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706378
Y. Xuan, Ling-yu Duan, Tiejun Huang
Coming with the rapid spread of Intelligent terminals with camera, mobile visual search techniques have undergone a revolution, where visual information can be easily browsed and retrieved upon simply capturing a query photo. However, most existing work targets at compact description of natural scene image statistics, while dealing with line drawing images retains an open problem. This paper presents a unified framework of line drawing problems in mobile visual search. We propose a compact description of line drawing image named Local Inner-Distance Shape Context (LISC) which is robust to the distortion and occlusion and enjoys scale and rotation invariance. Together with an innovative compression scheme using JBIG2 to reduce query delivery latency, our framework works well on both a self-built dataset and MPEG- 7 CE Shape-1 dataset. Promising results on both datasets show significant improvement over state-of-the-art algorithms.
随着带摄像头的智能终端的迅速普及,移动视觉搜索技术发生了一场革命,只需拍摄一张查询照片就可以轻松浏览和检索视觉信息。然而,大多数现有的工作都是针对自然场景图像统计的紧凑描述,而处理线条绘制图像仍然是一个开放的问题。本文提出了移动视觉搜索中线条绘制问题的统一框架。我们提出了一种紧凑的线条图像描述方法,称为局部内距离形状上下文(LISC),该方法对扭曲和遮挡具有鲁棒性,并且具有尺度和旋转不变性。结合使用JBIG2的创新压缩方案来减少查询交付延迟,我们的框架在自建数据集和MPEG- 7 CE Shape-1数据集上都能很好地工作。在这两个数据集上的令人鼓舞的结果表明,与最先进的算法相比,有了显著的改进。
{"title":"A local shape descriptor for mobile linedrawing retrieval","authors":"Y. Xuan, Ling-yu Duan, Tiejun Huang","doi":"10.1109/VCIP.2013.6706378","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706378","url":null,"abstract":"Coming with the rapid spread of Intelligent terminals with camera, mobile visual search techniques have undergone a revolution, where visual information can be easily browsed and retrieved upon simply capturing a query photo. However, most existing work targets at compact description of natural scene image statistics, while dealing with line drawing images retains an open problem. This paper presents a unified framework of line drawing problems in mobile visual search. We propose a compact description of line drawing image named Local Inner-Distance Shape Context (LISC) which is robust to the distortion and occlusion and enjoys scale and rotation invariance. Together with an innovative compression scheme using JBIG2 to reduce query delivery latency, our framework works well on both a self-built dataset and MPEG- 7 CE Shape-1 dataset. Promising results on both datasets show significant improvement over state-of-the-art algorithms.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"1630 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129265165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quality enhancement based on retinex and pseudo-HDR synthesis algorithms for endoscopic images 基于retinex和伪hdr合成算法的内镜图像质量增强
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706375
J. Wu, Guo-Shiang Lin, Hsiao-Ting Hsu, You-Peng Liao, Kai-Che Liu, W. Lie
In this paper, we present a quality enhancement scheme for endoscopic images. Traditional algorithms might be able to enhance the image contrast, but possible over-enhancement also lead to bad overall visual quality which prevents surgeons from accurate examination or operations of instruments in Minimal Invasive Surgery (MIS). Our proposed scheme integrates the well-known retinex algorithm with a pseudo-HDR (High Dynamic Range) synthesis process, designed to compose of three parts: multiscale retinex with gamma correction (MSR-G), local brightness range expansion (brightness diversity), and bilateral-filter-based HDR image fusion. Experiment results demonstrate that the proposed scheme is able to enhance image details and keep the overall visual quality good as well, with respect to other existing methods.
在本文中,我们提出了一种内镜图像的质量增强方案。传统的算法可以增强图像对比度,但可能的过度增强也会导致整体视觉质量差,妨碍外科医生在微创手术(MIS)中准确检查或操作器械。我们提出的方案将著名的retinex算法与伪HDR(高动态范围)合成过程相结合,设计由三个部分组成:带伽马校正的多尺度retinex (MSR-G)、局部亮度范围扩展(亮度多样性)和基于双边滤波器的HDR图像融合。实验结果表明,与现有方法相比,该方法在增强图像细节的同时,还能保持较好的整体视觉质量。
{"title":"Quality enhancement based on retinex and pseudo-HDR synthesis algorithms for endoscopic images","authors":"J. Wu, Guo-Shiang Lin, Hsiao-Ting Hsu, You-Peng Liao, Kai-Che Liu, W. Lie","doi":"10.1109/VCIP.2013.6706375","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706375","url":null,"abstract":"In this paper, we present a quality enhancement scheme for endoscopic images. Traditional algorithms might be able to enhance the image contrast, but possible over-enhancement also lead to bad overall visual quality which prevents surgeons from accurate examination or operations of instruments in Minimal Invasive Surgery (MIS). Our proposed scheme integrates the well-known retinex algorithm with a pseudo-HDR (High Dynamic Range) synthesis process, designed to compose of three parts: multiscale retinex with gamma correction (MSR-G), local brightness range expansion (brightness diversity), and bilateral-filter-based HDR image fusion. Experiment results demonstrate that the proposed scheme is able to enhance image details and keep the overall visual quality good as well, with respect to other existing methods.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132380689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Correlation estimation for distributed wireless video communication 分布式无线视频通信的相关估计
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706372
Xiaoliang Zhu, N. Zhang, Xiaopeng Fan, Ruiqin Xiong, Debin Zhao
One important problem in distributed video coding is to estimate the variance of the correlation noise between the video signal and its decoder side information. This variance is hard to estimate due to the lack of the motion vectors at the encoder side. In this paper, we first propose a linear model to estimate this variance by referring the zero motion prediction at the encoder based on a Markov field assumption. Furthermore, not only the prediction noise from the video signal itself but also the additional noise due to wireless transmission is considered in this paper. We applied our correlation estimation method in our recent distributed wireless visual communication framework called DCAST. The experimental results show that the proposed method improves the video PSNR by 0.5-1.5dB while avoiding motion estimation at encoder.
分布式视频编码的一个重要问题是估计视频信号与其解码器侧信息之间的相关噪声的方差。由于编码器侧缺乏运动向量,这种方差很难估计。在本文中,我们首先提出了一个线性模型来估计这个方差参考零运动预测在编码器基于马尔可夫场假设。此外,本文不仅考虑了视频信号本身的预测噪声,还考虑了由于无线传输而产生的附加噪声。我们将相关估计方法应用到最新的分布式无线视觉通信框架DCAST中。实验结果表明,该方法在避免编码器运动估计的情况下,将视频的PSNR提高了0.5 ~ 1.5 db。
{"title":"Correlation estimation for distributed wireless video communication","authors":"Xiaoliang Zhu, N. Zhang, Xiaopeng Fan, Ruiqin Xiong, Debin Zhao","doi":"10.1109/VCIP.2013.6706372","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706372","url":null,"abstract":"One important problem in distributed video coding is to estimate the variance of the correlation noise between the video signal and its decoder side information. This variance is hard to estimate due to the lack of the motion vectors at the encoder side. In this paper, we first propose a linear model to estimate this variance by referring the zero motion prediction at the encoder based on a Markov field assumption. Furthermore, not only the prediction noise from the video signal itself but also the additional noise due to wireless transmission is considered in this paper. We applied our correlation estimation method in our recent distributed wireless visual communication framework called DCAST. The experimental results show that the proposed method improves the video PSNR by 0.5-1.5dB while avoiding motion estimation at encoder.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"148 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123455940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ultra high-definition video coding using bit-depth reduction with image noise reduction and pseudo-contour prevention 超高清视频编码采用位深降低与图像降噪和伪轮廓预防
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706360
Y. Matsuo, T. Misu, S. Iwamura, S. Sakaida
We propose a novel ultra high-definition video coding method with bit-depth reduction before encoding procedure and bit-depth reconstruction after decoding procedure. The bit-depth reduction is performed by Lloyd-Max quantization; considering ultra high-definition video noise reduction for high coding efficiency and gradation conservation for pseudo-contour prevention. The bit-depth reconstruction is carried out accurately using side information which is determined by comparing a local-decoded bit-depth reconstructed image and an original image on encoder side. Experiments show that the proposed method has a pseudo-contour prevention effect and a better PSNR in comparison with conventional video coding methods.
提出了一种新的超高清视频编码方法,在编码前进行位深降低,在解码后进行位深重建。比特深度缩减采用Lloyd-Max量化;考虑超高清视频降噪以提高编码效率,考虑渐变守恒以防止伪轮廓。通过比较局部解码后的位深度重构图像与编码器侧的原始图像,利用侧信息精确地进行位深度重构。实验表明,与传统的视频编码方法相比,该方法具有伪轮廓预防效果和更好的PSNR。
{"title":"Ultra high-definition video coding using bit-depth reduction with image noise reduction and pseudo-contour prevention","authors":"Y. Matsuo, T. Misu, S. Iwamura, S. Sakaida","doi":"10.1109/VCIP.2013.6706360","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706360","url":null,"abstract":"We propose a novel ultra high-definition video coding method with bit-depth reduction before encoding procedure and bit-depth reconstruction after decoding procedure. The bit-depth reduction is performed by Lloyd-Max quantization; considering ultra high-definition video noise reduction for high coding efficiency and gradation conservation for pseudo-contour prevention. The bit-depth reconstruction is carried out accurately using side information which is determined by comparing a local-decoded bit-depth reconstructed image and an original image on encoder side. Experiments show that the proposed method has a pseudo-contour prevention effect and a better PSNR in comparison with conventional video coding methods.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"223 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123724341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2013 Visual Communications and Image Processing (VCIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1