首页 > 最新文献

28th Picture Coding Symposium最新文献

英文 中文
Temporal signal energy correction and low-complexity encoder feedback for lossy scalable video coding 时间信号能量校正和低复杂度编码器反馈的有损可扩展视频编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702518
Marijn J. H. Loomans, Cornelis J. Koeleman, P. D. With
In this paper, we address two problems found in embedded implementations of Scalable Video Codecs (SVCs): the temporal signal energy distribution and frame-to-frame quality fluctuations. The unequal energy distribution between the low- and high-pass band with integer-based wavelets leads to sub-optimal rate-distortion choices coupled with quantization-error accumulations. The second problem is the quality fluctuation between frames within a Group Of Pictures (GOP). To solve these two problems, we present two modifications to the SVC. The first modification aims at a temporal energy correction of the lifting scheme in the temporal wavelet decomposition. By moving this energy correction to the leaves of the temporal tree, we can save on required memory size, bandwidth and computations, while reducing floating/fixed-point conversion errors. The second modification feeds back the decoded first frame of the GOP (the temporal low-pass) into the temporal coding chain. The decoding of the first frame is achieved without entropy decoding while avoiding any required modifications at the decoder. Experiments show that quality fluctuations within the GOP are significantly reduced, thereby significantly increasing the subjective visual quality. On top of this, a small quality improvement is achieved on average.
在本文中,我们解决了可扩展视频编解码器(SVCs)嵌入式实现中的两个问题:时间信号能量分布和帧间质量波动。基于整数的小波在低通和高通频带之间的能量分布不均匀,导致次优的率失真选择加上量化误差累积。第二个问题是一组图像(GOP)中帧之间的质量波动。为了解决这两个问题,我们对SVC进行了两种修改。第一个改进的目的是对时间小波分解中的提升方案进行时间能量校正。通过将这种能量校正移动到时间树的叶子,我们可以节省所需的内存大小、带宽和计算,同时减少浮点/定点转换错误。第二种修改将解码后的GOP第一帧(时序低通)反馈到时序编码链中。在没有熵解码的情况下实现第一帧的解码,同时避免在解码器处进行任何必要的修改。实验表明,显著降低了GOP内的质量波动,从而显著提高了主观视觉质量。在此基础上,平均实现了一个小的质量改进。
{"title":"Temporal signal energy correction and low-complexity encoder feedback for lossy scalable video coding","authors":"Marijn J. H. Loomans, Cornelis J. Koeleman, P. D. With","doi":"10.1109/PCS.2010.5702518","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702518","url":null,"abstract":"In this paper, we address two problems found in embedded implementations of Scalable Video Codecs (SVCs): the temporal signal energy distribution and frame-to-frame quality fluctuations. The unequal energy distribution between the low- and high-pass band with integer-based wavelets leads to sub-optimal rate-distortion choices coupled with quantization-error accumulations. The second problem is the quality fluctuation between frames within a Group Of Pictures (GOP). To solve these two problems, we present two modifications to the SVC. The first modification aims at a temporal energy correction of the lifting scheme in the temporal wavelet decomposition. By moving this energy correction to the leaves of the temporal tree, we can save on required memory size, bandwidth and computations, while reducing floating/fixed-point conversion errors. The second modification feeds back the decoded first frame of the GOP (the temporal low-pass) into the temporal coding chain. The decoding of the first frame is achieved without entropy decoding while avoiding any required modifications at the decoder. Experiments show that quality fluctuations within the GOP are significantly reduced, thereby significantly increasing the subjective visual quality. On top of this, a small quality improvement is achieved on average.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116295053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast rate-distortion optimized transform for Intra coding 快速率失真优化变换的Intra编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702552
Xin Zhao, Li Zhang, Siwei Ma, Wen Gao
In our previous work, the rate-distortion optimized transform (RDOT) is introduced for Intra coding, which is featured by the usage of multiple offline-trained transform matrix candidates. The proposed RDOT achieves remarkable coding gain for KTA Intra coding, while maintaining almost the same computational complexity at the decoder. However, at the encoder, the computational complexity is increased drastically by the expensive ratedistortion (R-D) optimized selection of transform matrix. To resolve this problem, in this paper, we propose a fast RDOT scheme using macroblock- and block-level R-D cost thresholding. With the proposed method, unnecessary mode trials and R-D evaluations of transform matrices can be efficiently skipped from the mode decision process. Extensive experimental results show that, with negligible coding performance degradation, about 88.9% of the total encoding time is saved by the proposed method.
在我们之前的工作中,引入了用于Intra编码的率失真优化变换(RDOT),其特点是使用多个离线训练的变换矩阵候选者。提出的RDOT在KTA Intra编码中实现了显著的编码增益,同时在解码器上保持了几乎相同的计算复杂度。然而,在编码器上,由于对变换矩阵进行了昂贵的率失真(R-D)优化选择,计算复杂度急剧增加。为了解决这一问题,本文提出了一种使用宏块和块级R-D成本阈值的快速RDOT方案。利用该方法,可以有效地跳过模式决策过程中不必要的模式试验和变换矩阵的R-D评估。大量的实验结果表明,在编码性能下降可以忽略不计的情况下,该方法节省了约88.9%的总编码时间。
{"title":"Fast rate-distortion optimized transform for Intra coding","authors":"Xin Zhao, Li Zhang, Siwei Ma, Wen Gao","doi":"10.1109/PCS.2010.5702552","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702552","url":null,"abstract":"In our previous work, the rate-distortion optimized transform (RDOT) is introduced for Intra coding, which is featured by the usage of multiple offline-trained transform matrix candidates. The proposed RDOT achieves remarkable coding gain for KTA Intra coding, while maintaining almost the same computational complexity at the decoder. However, at the encoder, the computational complexity is increased drastically by the expensive ratedistortion (R-D) optimized selection of transform matrix. To resolve this problem, in this paper, we propose a fast RDOT scheme using macroblock- and block-level R-D cost thresholding. With the proposed method, unnecessary mode trials and R-D evaluations of transform matrices can be efficiently skipped from the mode decision process. Extensive experimental results show that, with negligible coding performance degradation, about 88.9% of the total encoding time is saved by the proposed method.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"107 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125471426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Suppressing texture-depth misalignment for boundary noise removal in view synthesis 抑制纹理深度偏差,消除视图合成中的边界噪声
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702494
Yin Zhao, Zhenzhong Chen, Dong Tian, Ce Zhu, Lu Yu
During view synthesis based on depth maps, also known as Depth-Image-Based Rendering (DIBR), annoying artifacts are often generated around foreground objects, yielding the visual effects that slim silhouettes of foreground objects are scattered into the background. The artifacts are referred as the boundary noises. We investigate the cause of boundary noises, and find out that they result from the misalignment between texture and depth information along object boundaries. Accordingly, we propose a novel solution to remove such boundary noises by applying restrictions during forward warping on the pixels within the texture-depth misalignment regions. Experiments show this algorithm can effectively eliminate most boundary noises and it is also robust for view synthesis with compressed depth and texture information.
在基于深度图的视图合成过程中,也称为基于深度图像的渲染(deep - image - based Rendering, DIBR),前景物体周围经常会产生令人讨厌的伪影,从而产生前景物体的细长轮廓分散到背景中的视觉效果。这些伪影被称为边界噪声。研究了边界噪声产生的原因,发现边界噪声的产生是由于物体边界上纹理和深度信息不一致造成的。因此,我们提出了一种新的解决方案,通过对纹理深度不对齐区域内的像素在前向翘曲期间施加限制来去除这种边界噪声。实验表明,该算法能够有效地消除大部分边界噪声,对压缩深度和纹理信息的图像合成具有较强的鲁棒性。
{"title":"Suppressing texture-depth misalignment for boundary noise removal in view synthesis","authors":"Yin Zhao, Zhenzhong Chen, Dong Tian, Ce Zhu, Lu Yu","doi":"10.1109/PCS.2010.5702494","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702494","url":null,"abstract":"During view synthesis based on depth maps, also known as Depth-Image-Based Rendering (DIBR), annoying artifacts are often generated around foreground objects, yielding the visual effects that slim silhouettes of foreground objects are scattered into the background. The artifacts are referred as the boundary noises. We investigate the cause of boundary noises, and find out that they result from the misalignment between texture and depth information along object boundaries. Accordingly, we propose a novel solution to remove such boundary noises by applying restrictions during forward warping on the pixels within the texture-depth misalignment regions. Experiments show this algorithm can effectively eliminate most boundary noises and it is also robust for view synthesis with compressed depth and texture information.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126224952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Colorization-based coding by focusing on characteristics of colorization bases 基于着色的编码,重点关注着色基的特征
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702473
Shunsuke Ono, T. Miyata, Y. Sakai
Colorization is a method that adds color components to a grayscale image using only a few representative pixels provided by the user. A novel approach to image compression called colorization-based coding has recently been proposed. It automatically extracts representative pixels from an original color image at an encoder and restores a full color image by using colorization at a decoder. However, previous studies on colorization-based coding extract redundant representative pixels and do not extract the pixels required for suppressing coding error. This paper focuses on the colorization basis that restricts the decoded color components. From this viewpoint, we propose a new colorization-based coding method. Experimental results revealed that our method can drastically suppress the information amount (number of representative pixels) compared conventional colorization based-coding while objective quality is maintained.
着色是一种仅使用用户提供的几个代表性像素向灰度图像添加颜色组件的方法。最近提出了一种新的图像压缩方法,称为基于颜色的编码。它在编码器处自动从原始彩色图像中提取有代表性的像素,并通过在解码器处使用着色来恢复全彩色图像。然而,以往的基于颜色的编码研究提取了冗余的代表性像素,而没有提取抑制编码误差所需的像素。本文重点研究了限制解码颜色分量的着色基础。从这个角度出发,我们提出了一种新的基于着色的编码方法。实验结果表明,与传统的基于颜色的编码相比,我们的方法可以在保持客观质量的同时显著抑制信息量(代表像素数)。
{"title":"Colorization-based coding by focusing on characteristics of colorization bases","authors":"Shunsuke Ono, T. Miyata, Y. Sakai","doi":"10.1109/PCS.2010.5702473","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702473","url":null,"abstract":"Colorization is a method that adds color components to a grayscale image using only a few representative pixels provided by the user. A novel approach to image compression called colorization-based coding has recently been proposed. It automatically extracts representative pixels from an original color image at an encoder and restores a full color image by using colorization at a decoder. However, previous studies on colorization-based coding extract redundant representative pixels and do not extract the pixels required for suppressing coding error. This paper focuses on the colorization basis that restricts the decoded color components. From this viewpoint, we propose a new colorization-based coding method. Experimental results revealed that our method can drastically suppress the information amount (number of representative pixels) compared conventional colorization based-coding while objective quality is maintained.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121800821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Entropy coding in video compression using probability interval partitioning 基于概率间隔分割的视频压缩熵编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702580
D. Marpe, H. Schwarz, T. Wiegand
We present a novel approach to entropy coding, which provides the coding efficiency and simple probability modeling capability of arithmetic coding at the complexity level of Huffman coding. The key element of the proposed approach is a partitioning of the unit interval into a small set of probability intervals. An input sequence of discrete source symbols is mapped to a sequence of binary symbols and each of the binary symbols is assigned to one of the probability intervals. The binary symbols that are assigned to a particular probability interval are coded at a fixed probability using a simple code that maps a variable number of binary symbols to variable length codewords. The probability modeling is decoupled from the actual binary entropy coding. The coding efficiency of the probability interval partitioning entropy (PIPE) coding is comparable to that of arithmetic coding.
我们提出了一种新的熵编码方法,它在霍夫曼编码的复杂性水平上提供了算术编码的编码效率和简单的概率建模能力。该方法的关键要素是将单位区间划分为小概率区间集。将离散源符号的输入序列映射到二进制符号序列,并将每个二进制符号分配到其中一个概率区间。分配给特定概率区间的二进制符号以固定概率编码,使用简单代码将可变数量的二进制符号映射到可变长度的码字。概率建模与实际的二值熵编码解耦。概率区间划分熵(PIPE)编码的编码效率与算术编码相当。
{"title":"Entropy coding in video compression using probability interval partitioning","authors":"D. Marpe, H. Schwarz, T. Wiegand","doi":"10.1109/PCS.2010.5702580","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702580","url":null,"abstract":"We present a novel approach to entropy coding, which provides the coding efficiency and simple probability modeling capability of arithmetic coding at the complexity level of Huffman coding. The key element of the proposed approach is a partitioning of the unit interval into a small set of probability intervals. An input sequence of discrete source symbols is mapped to a sequence of binary symbols and each of the binary symbols is assigned to one of the probability intervals. The binary symbols that are assigned to a particular probability interval are coded at a fixed probability using a simple code that maps a variable number of binary symbols to variable length codewords. The probability modeling is decoupled from the actual binary entropy coding. The coding efficiency of the probability interval partitioning entropy (PIPE) coding is comparable to that of arithmetic coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128274422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Parallel processing method for realtime FTV 实时FTV的并行处理方法
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702500
Kazuma Suzuki, Norishige Fukushima, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto
In this paper, we propose a parallel processing method to generate free viewpoint image in realtime. It is impossible to arrange the cameras in a high density realistically though it is necessary to capture images of the scene from innumerable cameras to express the free viewpoint image. Therefore, it is necessary to interpolate the image of arbitrary viewpoint from limited captured images. However, this process has the relation of the trade-off between the image quality and the computing time. In proposed method, it aimed to generate the high-quality free viewpoint image in realtime by applying the parallel processing to time-consuming interpolation part.
本文提出了一种实时生成自由视点图像的并行处理方法。为了表达自由视点图像,必须从无数台摄像机中捕捉现场图像,但要使摄像机真实地高密度排列是不可能的。因此,有必要从有限的捕获图像中插值任意视点的图像。然而,该过程存在图像质量与计算时间之间的权衡关系。该方法通过对耗时的插值部分进行并行处理,实时生成高质量的自由视点图像。
{"title":"Parallel processing method for realtime FTV","authors":"Kazuma Suzuki, Norishige Fukushima, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto","doi":"10.1109/PCS.2010.5702500","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702500","url":null,"abstract":"In this paper, we propose a parallel processing method to generate free viewpoint image in realtime. It is impossible to arrange the cameras in a high density realistically though it is necessary to capture images of the scene from innumerable cameras to express the free viewpoint image. Therefore, it is necessary to interpolate the image of arbitrary viewpoint from limited captured images. However, this process has the relation of the trade-off between the image quality and the computing time. In proposed method, it aimed to generate the high-quality free viewpoint image in realtime by applying the parallel processing to time-consuming interpolation part.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114988078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Stereoscopic depth estimation using fuzzy segment matching 基于模糊分段匹配的立体深度估计
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702443
K. Wegner, O. Stankiewicz, M. Domański
Stereo matching techniques usually match segments or blocks of pixels. This paper proposes to match segments defined as fuzzy sets of pixels. The proposed matching method is applicable to various techniques of stereo matching as well as to different measures of differences between pixels. In the paper, embedment of this approach into the state-of-the-art depth estimation software is described. Obtained experimental results show that the proposed way of stereo matching increases reliability of various depth estimation techniques.
立体匹配技术通常匹配像素段或像素块。本文提出了一种定义为模糊像素集的分段匹配方法。所提出的匹配方法适用于各种立体匹配技术,也适用于像素间差异的不同度量。本文描述了将这种方法嵌入到最先进的深度估计软件中。实验结果表明,所提出的立体匹配方法提高了各种深度估计技术的可靠性。
{"title":"Stereoscopic depth estimation using fuzzy segment matching","authors":"K. Wegner, O. Stankiewicz, M. Domański","doi":"10.1109/PCS.2010.5702443","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702443","url":null,"abstract":"Stereo matching techniques usually match segments or blocks of pixels. This paper proposes to match segments defined as fuzzy sets of pixels. The proposed matching method is applicable to various techniques of stereo matching as well as to different measures of differences between pixels. In the paper, embedment of this approach into the state-of-the-art depth estimation software is described. Obtained experimental results show that the proposed way of stereo matching increases reliability of various depth estimation techniques.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132455746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Technical design & IPR analysis for royalty-free video codecs 免版税视频编解码器的技术设计和知识产权分析
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702433
C. Reader
Royalty-free standards for image and video coding have been actively discussed for over 20 years. This paper breaks down the issues of designing royalty-free codecs into the major topics of requirements, video coding tools, classes of patents and performance. By dissecting the codec using a hierarchy of major to minor coding tools, it is possible to pinpoint where a patent impacts the video coding, and what the consequence will be of avoiding the patented tool.
图像和视频编码的免版税标准已经被积极讨论了20多年。本文将设计免版税编解码器的问题分解为需求、视频编码工具、专利类别和性能等主要主题。通过使用从主要到次要的编码工具的层次结构来剖析编解码器,可以确定专利对视频编码的影响,以及避免专利工具的后果。
{"title":"Technical design & IPR analysis for royalty-free video codecs","authors":"C. Reader","doi":"10.1109/PCS.2010.5702433","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702433","url":null,"abstract":"Royalty-free standards for image and video coding have been actively discussed for over 20 years. This paper breaks down the issues of designing royalty-free codecs into the major topics of requirements, video coding tools, classes of patents and performance. By dissecting the codec using a hierarchy of major to minor coding tools, it is possible to pinpoint where a patent impacts the video coding, and what the consequence will be of avoiding the patented tool.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131063036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Free-viewpoint image generation using different focal length camera array 利用不同焦距相机阵列生成自由视点图像
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702508
Kengo Ando, Norishige Fukushima, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto
The availability of multi-view images of a scene makes new and exciting applications possible, including Free-Viewpoint TV (FTV). FTV allows us to change viewpoint freely in a 3D world, where the virtual viewpoint images are synthesized by Image-Based Rendering (IBR). In this paper, we introduce a FTV depth estimation method for forward virtual viewpoints. Moreover, we introduce a view generation method by using a zoom camera in our camera setup to improve virtual viewpoint-ts' image quality. Simulation results confirm reduced error during depth estimation using our proposed method in comparison with conventional stereo matching scheme. We have demonstrated the improvement in image resolution of virtually moved forward camera using a zoom camera setup.
场景的多视图图像的可用性使新的和令人兴奋的应用成为可能,包括自由视点电视(FTV)。FTV允许我们在3D世界中自由变换视点,其中虚拟视点图像是通过基于图像的渲染(IBR)合成的。本文介绍了一种前向虚拟视点的FTV深度估计方法。此外,我们在相机设置中引入了一种使用变焦相机的视图生成方法,以提高虚拟视点的图像质量。仿真结果表明,与传统的立体匹配方法相比,该方法在深度估计时减小了误差。我们演示了使用变焦相机设置虚拟向前移动相机的图像分辨率的提高。
{"title":"Free-viewpoint image generation using different focal length camera array","authors":"Kengo Ando, Norishige Fukushima, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto","doi":"10.1109/PCS.2010.5702508","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702508","url":null,"abstract":"The availability of multi-view images of a scene makes new and exciting applications possible, including Free-Viewpoint TV (FTV). FTV allows us to change viewpoint freely in a 3D world, where the virtual viewpoint images are synthesized by Image-Based Rendering (IBR). In this paper, we introduce a FTV depth estimation method for forward virtual viewpoints. Moreover, we introduce a view generation method by using a zoom camera in our camera setup to improve virtual viewpoint-ts' image quality. Simulation results confirm reduced error during depth estimation using our proposed method in comparison with conventional stereo matching scheme. We have demonstrated the improvement in image resolution of virtually moved forward camera using a zoom camera setup.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134139835","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Bit-plane compressive sensing with Bayesian decoding for lossy compression 有损压缩的位平面压缩感知与贝叶斯解码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702577
Sz-Hsien Wu, Wen-Hsiao Peng, Tihao Chiang
This paper addresses the problem of reconstructing a com-pressively sampled sparse signal from its lossy and possibly insufficient measurements. The process involves estimations of sparsity pattern and sparse representation, for which we derived a vector estimator based on the Maximum a Posteriori Probability (MAP) rule. By making full use of signal prior knowledge, our scheme can use a measurement number close to sparsity to achieve perfect reconstruction. It also shows a much lower error probability of sparsity pattern than prior work, given insufficient measurements. To better recover the most significant part of the sparse representation, we further introduce the notion of bit-plane separation. When applied to image compression, the technique in combination with our MAP estimator shows promising results as compared to JPEG: the difference in compression ratio is seen to be within a factor of two, given the same decoded quality.
本文解决了从有损和可能不充分的测量中重建压缩采样稀疏信号的问题。该过程涉及稀疏模式和稀疏表示的估计,为此我们推导了一个基于最大后验概率(MAP)规则的向量估计器。通过充分利用信号先验知识,我们的方案可以使用接近稀疏度的测量数来实现完美的重构。在测量不充分的情况下,它还显示出稀疏模式的错误概率比以前的工作低得多。为了更好地恢复稀疏表示的最重要部分,我们进一步引入了位平面分离的概念。当应用于图像压缩时,与我们的MAP估计器相结合的技术显示出与JPEG相比有希望的结果:在给定相同的解码质量的情况下,压缩比的差异被认为在两倍之内。
{"title":"Bit-plane compressive sensing with Bayesian decoding for lossy compression","authors":"Sz-Hsien Wu, Wen-Hsiao Peng, Tihao Chiang","doi":"10.1109/PCS.2010.5702577","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702577","url":null,"abstract":"This paper addresses the problem of reconstructing a com-pressively sampled sparse signal from its lossy and possibly insufficient measurements. The process involves estimations of sparsity pattern and sparse representation, for which we derived a vector estimator based on the Maximum a Posteriori Probability (MAP) rule. By making full use of signal prior knowledge, our scheme can use a measurement number close to sparsity to achieve perfect reconstruction. It also shows a much lower error probability of sparsity pattern than prior work, given insufficient measurements. To better recover the most significant part of the sparse representation, we further introduce the notion of bit-plane separation. When applied to image compression, the technique in combination with our MAP estimator shows promising results as compared to JPEG: the difference in compression ratio is seen to be within a factor of two, given the same decoded quality.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132136959","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
28th Picture Coding Symposium
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1