首页 > 最新文献

2014 IEEE Visual Communications and Image Processing Conference最新文献

英文 中文
Efficient lossless coding of highpass bands from block-based motion compensated wavelet lifting using JPEG 2000 基于JPEG 2000的基于块的运动补偿小波提升的高效无损编码
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051590
Wolfgang Schnurrer, Tobias Tröger, T. Richter, Jürgen Seiler, André Kaup
Lossless image coding is a crucial task especially in the medical area, e.g., for volumes from Computed Tomography or Magnetic Resonance Tomography. Besides lossless coding, compensated wavelet lifting offers a scalable representation of such huge volumes. While compensation methods increase the details in the lowpass band, they also vary the characteristics of the wavelet coefficients, so an adaption of the coefficient coder should be considered. We propose a simple invertible extension for JPEG 2000 that can reduce the filesize for lossless coding of the highpass band by 0.8% on average with peak rate saving of 1.1%.
无损图像编码是一项至关重要的任务,特别是在医学领域,例如,从计算机断层扫描或磁共振断层扫描的体积。除了无损编码,补偿小波提升提供了一个可伸缩的表示如此巨大的体积。补偿方法在增加低通带细节的同时,也改变了小波系数的特性,因此需要考虑系数编码器的自适应。我们为JPEG 2000提出了一个简单的可逆扩展,它可以将高通带无损编码的文件大小平均减少0.8%,峰值速率节省1.1%。
{"title":"Efficient lossless coding of highpass bands from block-based motion compensated wavelet lifting using JPEG 2000","authors":"Wolfgang Schnurrer, Tobias Tröger, T. Richter, Jürgen Seiler, André Kaup","doi":"10.1109/VCIP.2014.7051590","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051590","url":null,"abstract":"Lossless image coding is a crucial task especially in the medical area, e.g., for volumes from Computed Tomography or Magnetic Resonance Tomography. Besides lossless coding, compensated wavelet lifting offers a scalable representation of such huge volumes. While compensation methods increase the details in the lowpass band, they also vary the characteristics of the wavelet coefficients, so an adaption of the coefficient coder should be considered. We propose a simple invertible extension for JPEG 2000 that can reduce the filesize for lossless coding of the highpass band by 0.8% on average with peak rate saving of 1.1%.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133804977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Foveation-based image quality assessment 基于注视点的图像质量评估
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051495
Wen-Jiin Tsai, Yi-Shih Liu
Since human vision has much greater resolutions at the center of our visual field than elsewhere, different criteria of quality assessment should be applied on the image areas with different visual resolutions. This paper proposed a foveation-based image quality assessment method which adopted different sizes of windows in quality assessment for a single image. Visual salience models which estimate visual attention regions are used to determine the foveation center and foveation resolution models are used to guide the selection of window sizes for the areas over spatial extent of the image. Finally, the quality scores obtained from different window sizes are pooled together to get a single value for the image. The proposed method has been applied to IQA metrics, SSIM, PSNR, and UQI. The result shows that both Spearman and Kendall correlation coefficients can be improved significantly by our foveation-based method.
由于人类视觉在视野中心的分辨率比其他地方要大得多,因此对于不同视觉分辨率的图像区域,应采用不同的质量评估标准。本文提出了一种基于注视点的图像质量评价方法,该方法采用不同大小的窗口对单幅图像进行质量评价。利用估计视觉注意区域的视觉显著性模型来确定注视中心,利用注视分辨率模型来指导图像空间范围内区域窗口大小的选择。最后,将从不同窗口大小获得的质量分数汇集在一起,以获得图像的单个值。该方法已应用于IQA指标、SSIM、PSNR和UQI。结果表明,基于注视点的方法可以显著提高Spearman和Kendall相关系数。
{"title":"Foveation-based image quality assessment","authors":"Wen-Jiin Tsai, Yi-Shih Liu","doi":"10.1109/VCIP.2014.7051495","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051495","url":null,"abstract":"Since human vision has much greater resolutions at the center of our visual field than elsewhere, different criteria of quality assessment should be applied on the image areas with different visual resolutions. This paper proposed a foveation-based image quality assessment method which adopted different sizes of windows in quality assessment for a single image. Visual salience models which estimate visual attention regions are used to determine the foveation center and foveation resolution models are used to guide the selection of window sizes for the areas over spatial extent of the image. Finally, the quality scores obtained from different window sizes are pooled together to get a single value for the image. The proposed method has been applied to IQA metrics, SSIM, PSNR, and UQI. The result shows that both Spearman and Kendall correlation coefficients can be improved significantly by our foveation-based method.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"25 12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132122505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Depth map super-resolution via iterative joint-trilateral-upsampling 基于迭代联合三边上采样的深度图超分辨率
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051587
Yangguang Li, Lei Zhang, Yongbing Zhang, Huiming Xuan, Qionghai Dai
In this paper, we propose a new approach to solve the depth map super-resolution (SR) and denoising problems simultaneously. Inspired by joint-bilateral-upsampling (JBU), we devised the joint-trilateral-upsampling (JTU), which takes edge of the initial depth map, texture of the corresponding high-resolution color image and the values of the surrounding depth pixels, into consideration during the process of SR. To preserve the sharp edge of the up-sampled depth map and remove the noise, we introduce an iterative implementation, where current up-sampled depth map is fed into the next iteration, to refine the filter coefficients of JTU. The iterative JTU presents a high performance at many aspects such as sharping edge, denoising and none texture copying, etc. To demonstrate the superiority of the proposed method, we carry out various experiments and show an across-the-board quality improvement by both of subjective and objective evaluations compared with previous state-of-art methods.
本文提出了一种同时解决深度图超分辨率和去噪问题的新方法。受联合双向上采样(joint-bilateral-upsampling, JBU)的启发,我们设计了联合三边上采样(joint-trilateral-upsampling, JTU)方法,该方法在sr过程中考虑了初始深度图的边缘、相应高分辨率彩色图像的纹理以及周围深度像素的值。为了保持上采样深度图的锐利边缘并去除噪声,我们引入了一种迭代实现方法,将当前上采样深度图馈送到下一次迭代中。对JTU的滤波系数进行细化。迭代JTU在边缘锐化、去噪和无纹理复制等方面表现出较高的性能。为了证明所提出的方法的优越性,我们进行了各种实验,并通过主观和客观评价与以前的最先进的方法相比,显示了全面的质量改进。
{"title":"Depth map super-resolution via iterative joint-trilateral-upsampling","authors":"Yangguang Li, Lei Zhang, Yongbing Zhang, Huiming Xuan, Qionghai Dai","doi":"10.1109/VCIP.2014.7051587","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051587","url":null,"abstract":"In this paper, we propose a new approach to solve the depth map super-resolution (SR) and denoising problems simultaneously. Inspired by joint-bilateral-upsampling (JBU), we devised the joint-trilateral-upsampling (JTU), which takes edge of the initial depth map, texture of the corresponding high-resolution color image and the values of the surrounding depth pixels, into consideration during the process of SR. To preserve the sharp edge of the up-sampled depth map and remove the noise, we introduce an iterative implementation, where current up-sampled depth map is fed into the next iteration, to refine the filter coefficients of JTU. The iterative JTU presents a high performance at many aspects such as sharping edge, denoising and none texture copying, etc. To demonstrate the superiority of the proposed method, we carry out various experiments and show an across-the-board quality improvement by both of subjective and objective evaluations compared with previous state-of-art methods.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114949321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
A single octave SIFT algorithm for image feature extraction in resource limited hardware systems 在资源有限的硬件系统中,一种用于图像特征提取的单倍频次SIFT算法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051542
N.P. Borg, C. J. Debono, D. Zammit-Mangion
With the availability and rapid advancement of low-cost, low-power, and high-performance processors, machine vision is gaining popularity in various fields, including that of autonomous navigation systems. Applying feature extraction techniques on the captured images provides rich information about the surrounding environment that can be used to accurately determine the position, velocity, and orientation of a vehicle. To extract these features in such an application, we developed the Single Octave Scale Invariant Feature Transform (Single Octave SIFT). This solution drastically reduces the computational load and memory bandwidth requirements while providing an accurate image-based terrain referenced navigation system for micro- and small-sized Unmanned Aerial Vehicles (UAVs). The Gaussian filtering and Keypoint extraction stages are the most computationally intensive parts of the Single Octave SIFT. The main focus of this paper is the design of this modified SIFT algorithm and the basic building blocks needed to implement these two stages within a low-cost, low-power and small footprint Xilinx Spartan-6 LX150 FPGA. Simulation results show that the number of memory accesses is reduced by 99.7% for Full-HD (1920×1080) images1. The operation cycles of the Gaussian filtering and Keypoint extraction stages are reduced by 90.2% and 95% respectively, compared with the single instruction multiple data (SIMD) architecture.
随着低成本、低功耗和高性能处理器的可用性和快速发展,机器视觉在包括自主导航系统在内的各个领域越来越受欢迎。对捕获的图像应用特征提取技术,可以提供有关周围环境的丰富信息,可用于准确确定车辆的位置、速度和方向。为了在这样的应用中提取这些特征,我们开发了单八度尺度不变特征变换(Single Octave SIFT)。该解决方案大大降低了计算负载和内存带宽要求,同时为微型和小型无人机(uav)提供了精确的基于图像的地形参考导航系统。高斯滤波和关键点提取阶段是单倍频程SIFT中计算量最大的部分。本文的主要重点是改进SIFT算法的设计,以及在低成本,低功耗和小占地的Xilinx Spartan-6 LX150 FPGA中实现这两个阶段所需的基本构建块。仿真结果表明,对于全高清(1920×1080)图像,内存访问次数减少了99.7% 1。与单指令多数据(SIMD)架构相比,高斯滤波和关键点提取阶段的操作周期分别减少了90.2%和95%。
{"title":"A single octave SIFT algorithm for image feature extraction in resource limited hardware systems","authors":"N.P. Borg, C. J. Debono, D. Zammit-Mangion","doi":"10.1109/VCIP.2014.7051542","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051542","url":null,"abstract":"With the availability and rapid advancement of low-cost, low-power, and high-performance processors, machine vision is gaining popularity in various fields, including that of autonomous navigation systems. Applying feature extraction techniques on the captured images provides rich information about the surrounding environment that can be used to accurately determine the position, velocity, and orientation of a vehicle. To extract these features in such an application, we developed the Single Octave Scale Invariant Feature Transform (Single Octave SIFT). This solution drastically reduces the computational load and memory bandwidth requirements while providing an accurate image-based terrain referenced navigation system for micro- and small-sized Unmanned Aerial Vehicles (UAVs). The Gaussian filtering and Keypoint extraction stages are the most computationally intensive parts of the Single Octave SIFT. The main focus of this paper is the design of this modified SIFT algorithm and the basic building blocks needed to implement these two stages within a low-cost, low-power and small footprint Xilinx Spartan-6 LX150 FPGA. Simulation results show that the number of memory accesses is reduced by 99.7% for Full-HD (1920×1080) images1. The operation cycles of the Gaussian filtering and Keypoint extraction stages are reduced by 90.2% and 95% respectively, compared with the single instruction multiple data (SIMD) architecture.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"199 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131857632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Sample adaptive offset filter hardware design for HEVC encoder HEVC编码器自适应失调滤波器硬件设计示例
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051563
Fabiane Rediess, R. Conceição, B. Zatt, M. Porto, L. Agostini
This work presents a hardware design for the Sample Adaptive Offset filter, which is an innovation brought by the new video coding standard HEVC. The architectures focus on the encoder side and include both classification methods used in SAO, the Band Offset and Edge Offset, and also the statistical calculations for the offset generation. The proposed architectures feature two sample buffers, classification units for both SAO types and the statistical collection unit. The architectures were described in VHDL and synthesized to an Altera Stratix V FPGA. The synthesis results show that the proposed architectures achieve 364MHz and are capable to process 44 QFHD (3840×2160) frames per second using 8,040 ALUTs of the target device hardware resources.
本文提出了一种基于新视频编码标准HEVC的采样自适应偏移滤波器的硬件设计。该体系结构侧重于编码器端,包括SAO中使用的两种分类方法,频带偏移和边缘偏移,以及偏移生成的统计计算。所建议的体系结构具有两个样本缓冲区、SAO类型的分类单元和统计收集单元。用VHDL描述了这些结构,并将其合成到Altera Stratix V FPGA上。综合结果表明,所提出的架构达到364MHz,能够在目标设备硬件资源的8040个alut上每秒处理44个QFHD (3840×2160)帧。
{"title":"Sample adaptive offset filter hardware design for HEVC encoder","authors":"Fabiane Rediess, R. Conceição, B. Zatt, M. Porto, L. Agostini","doi":"10.1109/VCIP.2014.7051563","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051563","url":null,"abstract":"This work presents a hardware design for the Sample Adaptive Offset filter, which is an innovation brought by the new video coding standard HEVC. The architectures focus on the encoder side and include both classification methods used in SAO, the Band Offset and Edge Offset, and also the statistical calculations for the offset generation. The proposed architectures feature two sample buffers, classification units for both SAO types and the statistical collection unit. The architectures were described in VHDL and synthesized to an Altera Stratix V FPGA. The synthesis results show that the proposed architectures achieve 364MHz and are capable to process 44 QFHD (3840×2160) frames per second using 8,040 ALUTs of the target device hardware resources.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129514389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Depth-map driven planar surfaces detection 深度图驱动的平面检测
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051619
Jin Zhi, T. Tillo, Fei Cheng
Planar surface is a common feature in man-made structure, thus accurate detection of planar surface can benefit the image/video segmentation and reconstruction and also the navigation system of robots. Since depth map represents the distance from one object to the capturing camera in a grey image, it also can represent the surface characteristics of the objects. So in this paper, we propose a novel Depth-map Driven Planar Surface Detection (DDPSD) method, where detection starts from "the most flat" seed patch on the depth map and uses dynamic threshold value and surface function to control the growing process. Compared with one of the popular planar surface detection algorithms, RANdom SAmples Consensus (RANSAC), the accuracy of the proposed method is obviously superior on typical indoor scenes. Moreover, semi-planar surfaces can be also successfully detected by the proposed method.
平面是人造结构的一个共同特征,因此对平面的准确检测对图像/视频的分割和重建以及机器人的导航系统都有很大的帮助。由于深度图在灰度图像中表示一个物体到捕捉相机的距离,因此它也可以表示物体的表面特征。因此,本文提出了一种新的深度图驱动的平面表面检测(DDPSD)方法,该方法从深度图上“最平坦”的种子块开始检测,并使用动态阈值和表面函数来控制生长过程。与目前流行的一种平面表面检测算法RANSAC (RANdom SAmples Consensus)相比,该方法在典型室内场景下的检测精度明显优于随机样本一致性算法。此外,该方法还可以成功地检测半平面表面。
{"title":"Depth-map driven planar surfaces detection","authors":"Jin Zhi, T. Tillo, Fei Cheng","doi":"10.1109/VCIP.2014.7051619","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051619","url":null,"abstract":"Planar surface is a common feature in man-made structure, thus accurate detection of planar surface can benefit the image/video segmentation and reconstruction and also the navigation system of robots. Since depth map represents the distance from one object to the capturing camera in a grey image, it also can represent the surface characteristics of the objects. So in this paper, we propose a novel Depth-map Driven Planar Surface Detection (DDPSD) method, where detection starts from \"the most flat\" seed patch on the depth map and uses dynamic threshold value and surface function to control the growing process. Compared with one of the popular planar surface detection algorithms, RANdom SAmples Consensus (RANSAC), the accuracy of the proposed method is obviously superior on typical indoor scenes. Moreover, semi-planar surfaces can be also successfully detected by the proposed method.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132565288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Automatic denoising parameter estimation using gradient histograms 使用梯度直方图的自动去噪参数估计
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051580
Tamara Seybold, F. Kuhn, Julian Habigt, Mark Hartenstein, W. Stechele
State-of-the-art denoising methods provide denoising results that can be considered close to optimal. The denoising methods usually have one or more parameters regulating denoising strength that can be adapted for a specific image. To obtain the optimal denoising result, the correct parameter setting is crucial. In this paper, we therefore propose a method that can automatically estimate the optimal parameter of a denoising algorithm. Our approach compares the gradient histogram of a denoised image to an estimated reference gradient histogram. The reference gradient histogram is estimated based on down- and upsampling of the noisy image, thus our method works without a reference and is image-adaptive. We evaluate our propsed down-/upsampling-based gradient histogram method (DUG) based on a subjective test with 20 participants. In the test data, we included images from both the Kodak data set and the more realistic ARRI data set and we used the state-of-the-art denoising method BM3D. Based on the test results we can show that the parameter estimated by our method is very close to the human perception. Despite being very fast and simple to implement, our method shows a lower error than all other suitable no-reference metrics we found.
最先进的去噪方法可以提供接近最佳的去噪结果。去噪方法通常具有一个或多个调节去噪强度的参数,该参数可适应于特定图像。为了获得最佳的去噪效果,正确的参数设置至关重要。因此,本文提出了一种自动估计去噪算法最优参数的方法。我们的方法将去噪图像的梯度直方图与估计的参考梯度直方图进行比较。参考梯度直方图的估计是基于噪声图像的下采样和上采样,因此我们的方法可以在没有参考的情况下工作,并且是图像自适应的。我们基于20名参与者的主观测试评估了我们提出的基于下/上采样的梯度直方图方法(DUG)。在测试数据中,我们包括来自柯达数据集和更真实的ARRI数据集的图像,我们使用了最先进的去噪方法BM3D。测试结果表明,该方法估计的参数与人类感知非常接近。尽管实现起来非常快速和简单,但我们的方法比我们发现的所有其他合适的无参考指标显示出更低的误差。
{"title":"Automatic denoising parameter estimation using gradient histograms","authors":"Tamara Seybold, F. Kuhn, Julian Habigt, Mark Hartenstein, W. Stechele","doi":"10.1109/VCIP.2014.7051580","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051580","url":null,"abstract":"State-of-the-art denoising methods provide denoising results that can be considered close to optimal. The denoising methods usually have one or more parameters regulating denoising strength that can be adapted for a specific image. To obtain the optimal denoising result, the correct parameter setting is crucial. In this paper, we therefore propose a method that can automatically estimate the optimal parameter of a denoising algorithm. Our approach compares the gradient histogram of a denoised image to an estimated reference gradient histogram. The reference gradient histogram is estimated based on down- and upsampling of the noisy image, thus our method works without a reference and is image-adaptive. We evaluate our propsed down-/upsampling-based gradient histogram method (DUG) based on a subjective test with 20 participants. In the test data, we included images from both the Kodak data set and the more realistic ARRI data set and we used the state-of-the-art denoising method BM3D. Based on the test results we can show that the parameter estimated by our method is very close to the human perception. Despite being very fast and simple to implement, our method shows a lower error than all other suitable no-reference metrics we found.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"205 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131693924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
D-mago: A novel visual entity for storing emotional feeling with visual imprint 一种新颖的视觉实体,用于储存带有视觉印记的情感
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051607
Suk kyu Lee, Mungyu Bae, A. Y. Chung, Hwangnam Kim
Many users want to preserve their visual record of the moment that they want to commemorate. Nonetheless, it is still challenging to remember the actual emotional feeling for that moment even by looking at the old picture. There are methods such as to tag or hide the message within the image. However, tradeoffs exist by attaching additional data for the former method and the quality of the image is degraded for the latter one. It is difficult to avoid these two tradeoffs. In this paper, we propose D-mago to preserve the moment to remember as an image, which is consists of the visual information and the emotional feeling without binding extra data or degrading the quality of the image. To further verify the benefit of our proposed algorithm, we conducted series of evaluation studies to see the effectiveness of the proposed scheme. The results indicate that D-mago overcomes the preceding tradeoffPs by maintaining PSNR above 40 dB.
许多用户希望保存他们想要纪念的时刻的视觉记录。然而,即使是看着旧照片,也很难记住当时的真实情绪。有一些方法可以标记或隐藏图像中的消息。然而,对于前一种方法,由于附加了额外的数据而存在折衷,而对于后一种方法,图像质量会下降。很难避免这两种权衡。在本文中,我们提出D-mago将记忆瞬间保存为图像,它由视觉信息和情感感受组成,而不会绑定额外的数据或降低图像质量。为了进一步验证我们所提出的算法的效益,我们进行了一系列的评估研究,以了解所提出方案的有效性。结果表明,D-mago通过将PSNR保持在40 dB以上来克服上述权衡。
{"title":"D-mago: A novel visual entity for storing emotional feeling with visual imprint","authors":"Suk kyu Lee, Mungyu Bae, A. Y. Chung, Hwangnam Kim","doi":"10.1109/VCIP.2014.7051607","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051607","url":null,"abstract":"Many users want to preserve their visual record of the moment that they want to commemorate. Nonetheless, it is still challenging to remember the actual emotional feeling for that moment even by looking at the old picture. There are methods such as to tag or hide the message within the image. However, tradeoffs exist by attaching additional data for the former method and the quality of the image is degraded for the latter one. It is difficult to avoid these two tradeoffs. In this paper, we propose D-mago to preserve the moment to remember as an image, which is consists of the visual information and the emotional feeling without binding extra data or degrading the quality of the image. To further verify the benefit of our proposed algorithm, we conducted series of evaluation studies to see the effectiveness of the proposed scheme. The results indicate that D-mago overcomes the preceding tradeoffPs by maintaining PSNR above 40 dB.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130940934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A complexity reduction algorithm for depth maps intra prediction on the 3D-HEVC 基于3D-HEVC的深度图内预测复杂度降低算法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051523
G. Sanchez, Mário Saldanha, Gabriel Balota, B. Zatt, M. Porto, L. Agostini
This paper proposes a complexity reduction algorithm for the depth maps intra prediction of the emerging 3D High Efficiency Video Coding standard (3D-HEVC). The 3D-HEVC introduces a new set of specific tools for the depth map coding that includes four Depth Modeling Modes (DMM) and these new features have inserted extra effort on the intra prediction. This extra effort is undesired and contributes to increasing the power consumption, which is a huge problem especially for embedded-systems. For this reason, this paper proposes a complexity reduction algorithm for the DMM 1, called Gradient-Based Mode One Filter (GMOF). This algorithm applies a filter to the borders of the encoded block and determines the best positions to evaluate the DMM 1, reducing the computational effort of DMM 1 process. Experimental analysis showed that GMOF is capable to achieve, in average, a complexity reduction of 9.8% on depth maps prediction, when evaluating under Common Test Conditions (CTC), with minor impacts on the quality of the synthesized views.
针对新兴的3D高效视频编码标准(3D- hevc),提出了一种深度图帧内预测的复杂度降低算法。3D-HEVC为深度图编码引入了一套新的特定工具,其中包括四种深度建模模式(DMM),这些新功能为图像内预测增加了额外的工作量。这种额外的工作是不希望的,并且会增加功耗,这是一个巨大的问题,特别是对于嵌入式系统。为此,本文提出了一种dmm1的复杂度降低算法,称为基于梯度的模式一滤波器(GMOF)。该算法对编码块的边界进行滤波,确定对DMM - 1进行评估的最佳位置,减少了DMM - 1过程的计算量。实验分析表明,在通用测试条件(Common Test Conditions, CTC)下评估时,GMOF能够将深度图预测的复杂性平均降低9.8%,对合成视图的质量影响较小。
{"title":"A complexity reduction algorithm for depth maps intra prediction on the 3D-HEVC","authors":"G. Sanchez, Mário Saldanha, Gabriel Balota, B. Zatt, M. Porto, L. Agostini","doi":"10.1109/VCIP.2014.7051523","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051523","url":null,"abstract":"This paper proposes a complexity reduction algorithm for the depth maps intra prediction of the emerging 3D High Efficiency Video Coding standard (3D-HEVC). The 3D-HEVC introduces a new set of specific tools for the depth map coding that includes four Depth Modeling Modes (DMM) and these new features have inserted extra effort on the intra prediction. This extra effort is undesired and contributes to increasing the power consumption, which is a huge problem especially for embedded-systems. For this reason, this paper proposes a complexity reduction algorithm for the DMM 1, called Gradient-Based Mode One Filter (GMOF). This algorithm applies a filter to the borders of the encoded block and determines the best positions to evaluate the DMM 1, reducing the computational effort of DMM 1 process. Experimental analysis showed that GMOF is capable to achieve, in average, a complexity reduction of 9.8% on depth maps prediction, when evaluating under Common Test Conditions (CTC), with minor impacts on the quality of the synthesized views.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130983834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
A novel distributed compressive video sensing based on hybrid sparse basis 一种基于混合稀疏基的分布式压缩视频感知方法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051569
Haifeng Dong, Bojin Zhuang, Fei Su, Zhicheng Zhao
Distributed compressive video sensing (DCVS) is a new emerging video codec that incorporates advantages of distributed video coding (DVC) and compressive sensing (CS). However, due to the absence of a good sparse basis, the DCVS does not achieve ideal compressing efficiency compared with the traditional video codec, such as MPEG-4, H.264, etc. This paper proposes a new hybrid sparse basis, which combines the image-block prediction and DCT basis. Adaptive block-based prediction is employed to learn block-prediction basis by exploiting temporal correlation among successive frames. Based on linear DCT basis and predicted basis, the hybrid sparse basis can achieve sparser representation with lower complexity. The experiment results indicate that the proposal outperforms the state-of-the-art DCVS schemes on both visual quality and average PSNR. In addition, an iterative fashion proposed in the decoder can enhance the sparsity of the hybrid sparse basis and improve the rate-distortion performance significantly.
分布式压缩视频感知(DCVS)是一种融合了分布式视频编码(DVC)和压缩感知(CS)优点的新型视频编解码器。然而,由于缺乏良好的稀疏基,与传统的视频编解码器如MPEG-4、H.264等相比,DCVS的压缩效率并不理想。本文提出了一种新的混合稀疏基,将图像块预测与DCT基相结合。采用自适应分块预测,利用连续帧之间的时间相关性学习分块预测基础。混合稀疏基在线性DCT基和预测基的基础上,以较低的复杂度实现更稀疏的表示。实验结果表明,该方案在视觉质量和平均PSNR方面都优于目前最先进的DCVS方案。此外,在解码器中提出了一种迭代方式,可以增强混合稀疏基的稀疏性,显著提高码率失真性能。
{"title":"A novel distributed compressive video sensing based on hybrid sparse basis","authors":"Haifeng Dong, Bojin Zhuang, Fei Su, Zhicheng Zhao","doi":"10.1109/VCIP.2014.7051569","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051569","url":null,"abstract":"Distributed compressive video sensing (DCVS) is a new emerging video codec that incorporates advantages of distributed video coding (DVC) and compressive sensing (CS). However, due to the absence of a good sparse basis, the DCVS does not achieve ideal compressing efficiency compared with the traditional video codec, such as MPEG-4, H.264, etc. This paper proposes a new hybrid sparse basis, which combines the image-block prediction and DCT basis. Adaptive block-based prediction is employed to learn block-prediction basis by exploiting temporal correlation among successive frames. Based on linear DCT basis and predicted basis, the hybrid sparse basis can achieve sparser representation with lower complexity. The experiment results indicate that the proposal outperforms the state-of-the-art DCVS schemes on both visual quality and average PSNR. In addition, an iterative fashion proposed in the decoder can enhance the sparsity of the hybrid sparse basis and improve the rate-distortion performance significantly.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128449057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2014 IEEE Visual Communications and Image Processing Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1