首页 > 最新文献

2013 Visual Communications and Image Processing (VCIP)最新文献

英文 中文
Shaking video synthesis for video stabilization performance assessment 用于视频防抖性能评价的抖动视频合成
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706422
Hui Qu, Li Song, Gengjian Xue
The goal of video stabilization is to remove the unwanted camera motion and obtain stable versions. Theoretically, a good stabilization algorithm should remove the unwanted motion without the loss of image qualities. However, due to the lack of ground-truth video frames, the accurate performance evaluation of different algorithms is hard. Most existing evaluation techniques usually synthesize stable videos from shaking ones, but they are not effective enough. Different from previous methods, in this paper we propose a novel method which synthesize shaking videos from stable frames. Based on the synthetic shaking videos, we perform preliminary video stabilization performance assessment on three stabilization algorithms. Our shaking video synthesis method can not only give a benchmark for full-reference video stabilization performance assessment, but also provide a basis for exploring the theoretical bound of video stabilization which may help to improve existing stabilization algorithms.
视频防抖的目的是消除不必要的摄像机运动,获得稳定的版本。从理论上讲,一个好的稳定算法应该在不损失图像质量的情况下消除不必要的运动。然而,由于缺乏真实视频帧,很难对不同算法进行准确的性能评估。现有的评价技术大多是由震动视频合成稳定视频,但效果不理想。与以往的方法不同,本文提出了一种从稳定帧合成抖动视频的新方法。在合成震动视频的基础上,对三种稳定算法进行了初步的视频稳定性能评价。本文提出的抖动视频合成方法不仅可以为全参考视频防抖性能评估提供基准,而且可以为探索视频防抖的理论边界提供依据,有助于改进现有的防抖算法。
{"title":"Shaking video synthesis for video stabilization performance assessment","authors":"Hui Qu, Li Song, Gengjian Xue","doi":"10.1109/VCIP.2013.6706422","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706422","url":null,"abstract":"The goal of video stabilization is to remove the unwanted camera motion and obtain stable versions. Theoretically, a good stabilization algorithm should remove the unwanted motion without the loss of image qualities. However, due to the lack of ground-truth video frames, the accurate performance evaluation of different algorithms is hard. Most existing evaluation techniques usually synthesize stable videos from shaking ones, but they are not effective enough. Different from previous methods, in this paper we propose a novel method which synthesize shaking videos from stable frames. Based on the synthetic shaking videos, we perform preliminary video stabilization performance assessment on three stabilization algorithms. Our shaking video synthesis method can not only give a benchmark for full-reference video stabilization performance assessment, but also provide a basis for exploring the theoretical bound of video stabilization which may help to improve existing stabilization algorithms.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122225887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Long-term background memory based on Gaussian mixture model 基于高斯混合模型的长期背景记忆
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706397
W. Zhao, X. D. Zhao, W. M. Liu, X. L. Tang
This paper aims to present a long-term background memory framework, which is capable of memorizing long period background in video and rapidly adapting to the changes of background. Based on Gaussian mixture model (GMM), this framework enables an accurate identification of long period background appearances and presents a perfect solution to numerous typical problems on foreground detection. The experimental results with various benchmark sequences quantitatively and qualitatively demonstrate that the proposed algorithm outperforms many GMM-based methods for foreground detection, as well as other representative approaches.
本文旨在提出一种长时间背景记忆框架,能够记忆视频中的长时间背景,并能快速适应背景的变化。该框架基于高斯混合模型(GMM),能够准确地识别长周期背景,为前景检测中的许多典型问题提供了完美的解决方案。各种基准序列的定量和定性实验结果表明,该算法在前景检测方面优于许多基于gmm的方法,以及其他代表性方法。
{"title":"Long-term background memory based on Gaussian mixture model","authors":"W. Zhao, X. D. Zhao, W. M. Liu, X. L. Tang","doi":"10.1109/VCIP.2013.6706397","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706397","url":null,"abstract":"This paper aims to present a long-term background memory framework, which is capable of memorizing long period background in video and rapidly adapting to the changes of background. Based on Gaussian mixture model (GMM), this framework enables an accurate identification of long period background appearances and presents a perfect solution to numerous typical problems on foreground detection. The experimental results with various benchmark sequences quantitatively and qualitatively demonstrate that the proposed algorithm outperforms many GMM-based methods for foreground detection, as well as other representative approaches.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117186549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Mid-level feature based local descriptor selection for image search 图像搜索中基于中级特征的局部描述符选择
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706455
S. Bucak, A. Saxena, Abhishek Nagar, Felix C. A. Fernandes, Kong-Posh Bhat
The objective in developing compact descriptors for visual image search is building an image retrieval system that works efficiently and effectively under bandwidth and memory constraints. Selecting local descriptors to be processed, and sending them to the server for matching is an integral part of such a system. One such image search and retrieval system is the Compact Descriptors for Visual Search (CDVS) standardization test model being developed by MPEG which has an efficient local descriptor selection criteria. However, all the existing selection parameters in CDVS are based on low-level features. In this paper, we propose two “mid-level” local descriptor selection criteria: Visual Meaning Score (VMS), and Visual Vocabulary Score (VVS) which can be seamlessly integrated into the existing CDVS framework. A mid-level criteria explicitly allows selection of local descriptors closer to a given set of images. Both VMS and VVS are based on visual words (patches) of images, and provide significant gains over the current CDVS standard in terms of matching accuracy, and have very low implementation cost.
开发用于视觉图像搜索的紧凑描述符的目标是建立一个在带宽和内存限制下高效工作的图像检索系统。选择要处理的局部描述符,并将它们发送到服务器进行匹配,是这样一个系统的组成部分。其中一个图像搜索和检索系统是由MPEG开发的压缩视觉搜索描述符(CDVS)标准化测试模型,该模型具有高效的局部描述符选择标准。然而,现有的cddvs选择参数都是基于底层特征的。在本文中,我们提出了两个“中级”局部描述符选择标准:视觉意义评分(VMS)和视觉词汇评分(VVS),这两个标准可以无缝集成到现有的cddvs框架中。中级标准明确地允许选择更接近给定图像集的局部描述符。VMS和VVS都是基于图像的视觉词(补丁),在匹配精度方面比目前的CDVS标准有显著提高,并且实现成本非常低。
{"title":"Mid-level feature based local descriptor selection for image search","authors":"S. Bucak, A. Saxena, Abhishek Nagar, Felix C. A. Fernandes, Kong-Posh Bhat","doi":"10.1109/VCIP.2013.6706455","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706455","url":null,"abstract":"The objective in developing compact descriptors for visual image search is building an image retrieval system that works efficiently and effectively under bandwidth and memory constraints. Selecting local descriptors to be processed, and sending them to the server for matching is an integral part of such a system. One such image search and retrieval system is the Compact Descriptors for Visual Search (CDVS) standardization test model being developed by MPEG which has an efficient local descriptor selection criteria. However, all the existing selection parameters in CDVS are based on low-level features. In this paper, we propose two “mid-level” local descriptor selection criteria: Visual Meaning Score (VMS), and Visual Vocabulary Score (VVS) which can be seamlessly integrated into the existing CDVS framework. A mid-level criteria explicitly allows selection of local descriptors closer to a given set of images. Both VMS and VVS are based on visual words (patches) of images, and provide significant gains over the current CDVS standard in terms of matching accuracy, and have very low implementation cost.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114875313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A parallel root-finding method for omnidirectional image unwrapping 面向全向图像展开的并行寻根方法
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706340
N. Chong, M. D. Wong, Y. Kho
The panoramic unwrapping of catadioptric omnidirectional view (COV) sensors have mostly relied on a precomputed mapping look-up table due to an expensive computational load that generally has its bottleneck occur at solving a sextic polynomial. However, this approach causes a limitation to the viewpoint dynamics as runtime modifications to the mapping values are not allowed in the implementation. In this paper, a parallel root-finding technique using Compute Unified Device Architecture (CUDA) platform is proposed. The proposed method enables on-the-fly computation of the mapping look-up table thus facilitate in a real-time viewpoint adjustable panoramic unwrapping. Experimental results showed that the proposed implementation incurred minimum computational load, and performed at 10.3 times and 2.3 times the speed of a current generation central processing unit (CPU) respectively on a single-core and multi-core environment.
反射全向视图(COV)传感器的全景展开主要依赖于预先计算的映射查找表,由于计算量大,其瓶颈通常出现在求解六次多项式时。然而,这种方法会对视点动态造成限制,因为在实现中不允许对映射值进行运行时修改。本文提出了一种基于计算统一设备架构(CUDA)平台的并行寻根技术。该方法实现了地图查找表的实时计算,实现了视点可调全景展开。实验结果表明,在单核和多核环境下,该实现的计算负荷最小,执行速度分别是当前一代中央处理器(CPU)的10.3倍和2.3倍。
{"title":"A parallel root-finding method for omnidirectional image unwrapping","authors":"N. Chong, M. D. Wong, Y. Kho","doi":"10.1109/VCIP.2013.6706340","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706340","url":null,"abstract":"The panoramic unwrapping of catadioptric omnidirectional view (COV) sensors have mostly relied on a precomputed mapping look-up table due to an expensive computational load that generally has its bottleneck occur at solving a sextic polynomial. However, this approach causes a limitation to the viewpoint dynamics as runtime modifications to the mapping values are not allowed in the implementation. In this paper, a parallel root-finding technique using Compute Unified Device Architecture (CUDA) platform is proposed. The proposed method enables on-the-fly computation of the mapping look-up table thus facilitate in a real-time viewpoint adjustable panoramic unwrapping. Experimental results showed that the proposed implementation incurred minimum computational load, and performed at 10.3 times and 2.3 times the speed of a current generation central processing unit (CPU) respectively on a single-core and multi-core environment.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115848525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Salient object detection in image sequences via spatial-temporal cue 基于时空线索的图像序列显著目标检测
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706438
Chuang Gan, Zengchang Qin, Jia Xu, T. Wan
Contemporary video search and categorization are non-trivial tasks due to the massively increasing amount and content variety of videos. We put forward the study of visual saliency models in video. Such a model is employed to identify salient objects from the image background. Starting from the observation that motion information in video often attracts more human attention compared to static images, we devise a region contrast based saliency detection model using spatial-temporal cues (RCST). We introduce and study four saliency principles to realize the RCST. This generalizes the previous static image for saliency computational model to video. We conduct experiments on a publicly available video segmentation database where our method significantly outperforms seven state-of-the-art methods with respect to PR curve, ROC curve and visual comparison.
由于视频数量和内容的大量增加,当代视频搜索和分类是一项非常重要的任务。提出了视频中视觉显著性模型的研究。该模型用于从图像背景中识别显著目标。从观察到视频中的运动信息往往比静态图像更能吸引人们的注意力开始,我们设计了一个基于区域对比度的显著性检测模型,该模型使用时空线索(RCST)。我们介绍并研究了实现RCST的四种显著性原则。将以往的静态图像显著性计算模型推广到视频中。我们在一个公开可用的视频分割数据库上进行实验,我们的方法在PR曲线、ROC曲线和视觉比较方面明显优于7种最先进的方法。
{"title":"Salient object detection in image sequences via spatial-temporal cue","authors":"Chuang Gan, Zengchang Qin, Jia Xu, T. Wan","doi":"10.1109/VCIP.2013.6706438","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706438","url":null,"abstract":"Contemporary video search and categorization are non-trivial tasks due to the massively increasing amount and content variety of videos. We put forward the study of visual saliency models in video. Such a model is employed to identify salient objects from the image background. Starting from the observation that motion information in video often attracts more human attention compared to static images, we devise a region contrast based saliency detection model using spatial-temporal cues (RCST). We introduce and study four saliency principles to realize the RCST. This generalizes the previous static image for saliency computational model to video. We conduct experiments on a publicly available video segmentation database where our method significantly outperforms seven state-of-the-art methods with respect to PR curve, ROC curve and visual comparison.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121120919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-scale video text detection based on corner and stroke width verification 基于角和笔画宽度验证的多尺度视频文本检测
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706387
Boyu Zhang, Jiafeng Liu, Xianglong Tang
Focusing on the video text detection, which is challenging and with wide potential applications, a novel stroke width feature is proposed and a system which detects text regions based on multi-scale corner detection is implemented in this paper. In our system, candidate text regions are generated by applying morphologic operation based on corner points detected in different scales, and non-text regions are filtered by combining proposed stroke width feature with some simple geometric properties. Moreover, there is a new multi-instance semi-supervised learning strategy being proposed in this paper considering the unknown contrast parameter in stroke width extraction. Experiments taken on video frames from different kinds of video shots prove that the proposed approach is both efficient and accurate for video text detection.
针对视频文本检测这一具有挑战性和广阔应用前景的问题,提出了一种新的笔画宽度特征,并实现了一种基于多尺度角点检测的文本区域检测系统。在我们的系统中,基于不同尺度检测到的角点进行形态学运算生成候选文本区域,并结合提出的笔画宽度特征和一些简单的几何属性来过滤非文本区域。此外,本文还提出了一种考虑笔画宽度提取中对比度参数未知的多实例半监督学习策略。对不同类型的视频帧进行了实验,证明了该方法对视频文本检测的有效性和准确性。
{"title":"Multi-scale video text detection based on corner and stroke width verification","authors":"Boyu Zhang, Jiafeng Liu, Xianglong Tang","doi":"10.1109/VCIP.2013.6706387","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706387","url":null,"abstract":"Focusing on the video text detection, which is challenging and with wide potential applications, a novel stroke width feature is proposed and a system which detects text regions based on multi-scale corner detection is implemented in this paper. In our system, candidate text regions are generated by applying morphologic operation based on corner points detected in different scales, and non-text regions are filtered by combining proposed stroke width feature with some simple geometric properties. Moreover, there is a new multi-instance semi-supervised learning strategy being proposed in this paper considering the unknown contrast parameter in stroke width extraction. Experiments taken on video frames from different kinds of video shots prove that the proposed approach is both efficient and accurate for video text detection.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121827915","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A local shape descriptor for mobile linedrawing retrieval 用于移动线条检索的局部形状描述符
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706378
Y. Xuan, Ling-yu Duan, Tiejun Huang
Coming with the rapid spread of Intelligent terminals with camera, mobile visual search techniques have undergone a revolution, where visual information can be easily browsed and retrieved upon simply capturing a query photo. However, most existing work targets at compact description of natural scene image statistics, while dealing with line drawing images retains an open problem. This paper presents a unified framework of line drawing problems in mobile visual search. We propose a compact description of line drawing image named Local Inner-Distance Shape Context (LISC) which is robust to the distortion and occlusion and enjoys scale and rotation invariance. Together with an innovative compression scheme using JBIG2 to reduce query delivery latency, our framework works well on both a self-built dataset and MPEG- 7 CE Shape-1 dataset. Promising results on both datasets show significant improvement over state-of-the-art algorithms.
随着带摄像头的智能终端的迅速普及,移动视觉搜索技术发生了一场革命,只需拍摄一张查询照片就可以轻松浏览和检索视觉信息。然而,大多数现有的工作都是针对自然场景图像统计的紧凑描述,而处理线条绘制图像仍然是一个开放的问题。本文提出了移动视觉搜索中线条绘制问题的统一框架。我们提出了一种紧凑的线条图像描述方法,称为局部内距离形状上下文(LISC),该方法对扭曲和遮挡具有鲁棒性,并且具有尺度和旋转不变性。结合使用JBIG2的创新压缩方案来减少查询交付延迟,我们的框架在自建数据集和MPEG- 7 CE Shape-1数据集上都能很好地工作。在这两个数据集上的令人鼓舞的结果表明,与最先进的算法相比,有了显著的改进。
{"title":"A local shape descriptor for mobile linedrawing retrieval","authors":"Y. Xuan, Ling-yu Duan, Tiejun Huang","doi":"10.1109/VCIP.2013.6706378","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706378","url":null,"abstract":"Coming with the rapid spread of Intelligent terminals with camera, mobile visual search techniques have undergone a revolution, where visual information can be easily browsed and retrieved upon simply capturing a query photo. However, most existing work targets at compact description of natural scene image statistics, while dealing with line drawing images retains an open problem. This paper presents a unified framework of line drawing problems in mobile visual search. We propose a compact description of line drawing image named Local Inner-Distance Shape Context (LISC) which is robust to the distortion and occlusion and enjoys scale and rotation invariance. Together with an innovative compression scheme using JBIG2 to reduce query delivery latency, our framework works well on both a self-built dataset and MPEG- 7 CE Shape-1 dataset. Promising results on both datasets show significant improvement over state-of-the-art algorithms.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"1630 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129265165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quality enhancement based on retinex and pseudo-HDR synthesis algorithms for endoscopic images 基于retinex和伪hdr合成算法的内镜图像质量增强
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706375
J. Wu, Guo-Shiang Lin, Hsiao-Ting Hsu, You-Peng Liao, Kai-Che Liu, W. Lie
In this paper, we present a quality enhancement scheme for endoscopic images. Traditional algorithms might be able to enhance the image contrast, but possible over-enhancement also lead to bad overall visual quality which prevents surgeons from accurate examination or operations of instruments in Minimal Invasive Surgery (MIS). Our proposed scheme integrates the well-known retinex algorithm with a pseudo-HDR (High Dynamic Range) synthesis process, designed to compose of three parts: multiscale retinex with gamma correction (MSR-G), local brightness range expansion (brightness diversity), and bilateral-filter-based HDR image fusion. Experiment results demonstrate that the proposed scheme is able to enhance image details and keep the overall visual quality good as well, with respect to other existing methods.
在本文中,我们提出了一种内镜图像的质量增强方案。传统的算法可以增强图像对比度,但可能的过度增强也会导致整体视觉质量差,妨碍外科医生在微创手术(MIS)中准确检查或操作器械。我们提出的方案将著名的retinex算法与伪HDR(高动态范围)合成过程相结合,设计由三个部分组成:带伽马校正的多尺度retinex (MSR-G)、局部亮度范围扩展(亮度多样性)和基于双边滤波器的HDR图像融合。实验结果表明,与现有方法相比,该方法在增强图像细节的同时,还能保持较好的整体视觉质量。
{"title":"Quality enhancement based on retinex and pseudo-HDR synthesis algorithms for endoscopic images","authors":"J. Wu, Guo-Shiang Lin, Hsiao-Ting Hsu, You-Peng Liao, Kai-Che Liu, W. Lie","doi":"10.1109/VCIP.2013.6706375","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706375","url":null,"abstract":"In this paper, we present a quality enhancement scheme for endoscopic images. Traditional algorithms might be able to enhance the image contrast, but possible over-enhancement also lead to bad overall visual quality which prevents surgeons from accurate examination or operations of instruments in Minimal Invasive Surgery (MIS). Our proposed scheme integrates the well-known retinex algorithm with a pseudo-HDR (High Dynamic Range) synthesis process, designed to compose of three parts: multiscale retinex with gamma correction (MSR-G), local brightness range expansion (brightness diversity), and bilateral-filter-based HDR image fusion. Experiment results demonstrate that the proposed scheme is able to enhance image details and keep the overall visual quality good as well, with respect to other existing methods.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132380689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Detection of salient objects in computer synthesized images based on object-level contrast 基于目标级对比度的计算机合成图像中显著目标的检测
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706362
L. Dong, Weisi Lin, Yuming Fang, Shiqian Wu, S. H. Soon
In this work, we propose a method to detect visually salient objects in computer synthesized images from 3D meshes. Different from existing detection methods on graphic saliency which compute saliency based on pixel-level contrast, the proposed method computes saliency by measuring object-level contrast of each object to the other objects in a rendered image. Given a synthesized image, the proposed method first extracts dominant colors from each object, and represents each object with the dominant color descriptor (DCD). Saliency is measured as the contrast between the DCD of the object and the DCDs of its surrounding objects. We evaluate the proposed method on a data set of computer rendered images, and the results show that the proposed method obtains much better performance compared with existing related methods.
在这项工作中,我们提出了一种从3D网格中检测计算机合成图像中视觉显著目标的方法。与现有图形显著性检测方法基于像素级对比度计算显著性不同,该方法通过测量渲染图像中每个对象与其他对象的对象级对比度来计算显著性。该方法首先从合成图像中提取各目标的主色,并用主色描述符(DCD)表示每个目标。显著性是用物体的DCD与其周围物体的DCD之间的对比度来衡量的。在计算机渲染图像数据集上对所提方法进行了评价,结果表明所提方法比现有的相关方法获得了更好的性能。
{"title":"Detection of salient objects in computer synthesized images based on object-level contrast","authors":"L. Dong, Weisi Lin, Yuming Fang, Shiqian Wu, S. H. Soon","doi":"10.1109/VCIP.2013.6706362","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706362","url":null,"abstract":"In this work, we propose a method to detect visually salient objects in computer synthesized images from 3D meshes. Different from existing detection methods on graphic saliency which compute saliency based on pixel-level contrast, the proposed method computes saliency by measuring object-level contrast of each object to the other objects in a rendered image. Given a synthesized image, the proposed method first extracts dominant colors from each object, and represents each object with the dominant color descriptor (DCD). Saliency is measured as the contrast between the DCD of the object and the DCDs of its surrounding objects. We evaluate the proposed method on a data set of computer rendered images, and the results show that the proposed method obtains much better performance compared with existing related methods.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123739467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An improved image compression scheme with an adaptive parameters set in encrypted domain 一种改进的图像压缩方案,在加密域设置自适应参数
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706408
Guochao Zhang, Shaohui Liu, F. Jiang, Debin Zhao, Wen Gao
A growing societal awareness about privacy and security push the development of signal processing techniques in the encrypted domain. Data compression in encrypted domain attracts much attention recently years due to its avoiding the leakage of data source during compression. This paper proposes an improved block-by-block compression scheme of encrypted image with flexible compression ratio. The original image is encrypted by permuting the blocks of the image and then permuting the pixels in the blocks. In the compression, pixels chosen randomly used as reference information, and remaining pixels are compressed by coset code. At the decoder side, side information (SI) which is generated by combining correlation among blocks and image restoration from partial random samples (IRPRS) is utilized to assist the decompression. Moreover, an adaptive system parameters selection method is also given in this paper. The experimental results show that the proposed method can achieve a better reconstructed result compared with the earlier method.
社会对隐私和安全意识的提高推动了加密领域信号处理技术的发展。加密领域的数据压缩因其在压缩过程中避免了数据源的泄漏而受到近年来的广泛关注。提出了一种改进的具有灵活压缩比的加密图像逐块压缩方案。通过排列图像的块,然后排列块中的像素,对原始图像进行加密。在压缩过程中,随机选择像素作为参考信息,剩余像素通过协集编码进行压缩。在解码器侧,利用块间相关和部分随机样本图像恢复相结合产生的侧信息(SI)辅助解压缩。此外,本文还提出了一种自适应的系统参数选择方法。实验结果表明,与之前的方法相比,该方法可以获得更好的重构结果。
{"title":"An improved image compression scheme with an adaptive parameters set in encrypted domain","authors":"Guochao Zhang, Shaohui Liu, F. Jiang, Debin Zhao, Wen Gao","doi":"10.1109/VCIP.2013.6706408","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706408","url":null,"abstract":"A growing societal awareness about privacy and security push the development of signal processing techniques in the encrypted domain. Data compression in encrypted domain attracts much attention recently years due to its avoiding the leakage of data source during compression. This paper proposes an improved block-by-block compression scheme of encrypted image with flexible compression ratio. The original image is encrypted by permuting the blocks of the image and then permuting the pixels in the blocks. In the compression, pixels chosen randomly used as reference information, and remaining pixels are compressed by coset code. At the decoder side, side information (SI) which is generated by combining correlation among blocks and image restoration from partial random samples (IRPRS) is utilized to assist the decompression. Moreover, an adaptive system parameters selection method is also given in this paper. The experimental results show that the proposed method can achieve a better reconstructed result compared with the earlier method.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129036646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2013 Visual Communications and Image Processing (VCIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1