首页 > 最新文献

28th Picture Coding Symposium最新文献

英文 中文
Subjective evaluation of Hierarchical B-Frames using Video-MUSHRA 基于Video-MUSHRA的分层b帧主观评价
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702532
H. M. Mohammed, Nikolaus Färber
Hierarchical B-Frames (HBF) has emerged as an efficient video coding tool in recent years. As shown in the literature, this approach results in excellent PSNR gains of >1 dB. However these PSNR gains are not sufficiently assessed in a scientific manner by subjective tests. Hence in this paper, we evaluate HBF coding pattern subjectively by using the MUSHRA test methodology. While MUSHRA is well established in audio coding research, its application to video is a novelty of this paper. We compare HBF with simple IPP coding pattern at either same PSNR or same bit rate. Our results indicate that, HBF gains are clearly subjectively perceptible. Hence, it can be shown that PSNR gains also correlate with a subjective gain. Interestingly, even at same PSNR, HBF is found to be subjectively superior to simple IPP coding.
分层b帧(HBF)是近年来出现的一种高效的视频编码工具。如文献所示,这种方法可以获得>1 dB的极好PSNR增益。然而,这些PSNR增益没有通过主观测试以科学的方式充分评估。因此,本文采用MUSHRA测试方法对HBF编码模式进行主观上的评价。虽然MUSHRA在音频编码研究中已经有了很好的应用,但它在视频编码中的应用是本文的一个新课题。我们比较了HBF和简单的IPP编码模式在相同的PSNR或相同的比特率。我们的结果表明,HBF的增益是明显的主观可感知的。因此,可以证明PSNR增益也与主观增益相关。有趣的是,即使在相同的PSNR下,HBF也被发现主观上优于简单的IPP编码。
{"title":"Subjective evaluation of Hierarchical B-Frames using Video-MUSHRA","authors":"H. M. Mohammed, Nikolaus Färber","doi":"10.1109/PCS.2010.5702532","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702532","url":null,"abstract":"Hierarchical B-Frames (HBF) has emerged as an efficient video coding tool in recent years. As shown in the literature, this approach results in excellent PSNR gains of >1 dB. However these PSNR gains are not sufficiently assessed in a scientific manner by subjective tests. Hence in this paper, we evaluate HBF coding pattern subjectively by using the MUSHRA test methodology. While MUSHRA is well established in audio coding research, its application to video is a novelty of this paper. We compare HBF with simple IPP coding pattern at either same PSNR or same bit rate. Our results indicate that, HBF gains are clearly subjectively perceptible. Hence, it can be shown that PSNR gains also correlate with a subjective gain. Interestingly, even at same PSNR, HBF is found to be subjectively superior to simple IPP coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121406101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Decoder-side hierarchical motion estimation for dense vector fields 密集向量场解码器侧分层运动估计
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702509
S. Klomp, Marco Munderloh, J. Ostermann
Current video coding standards perform motion estimation at the encoder to predict frames prior to coding them. Since the decoder does not possess the source frames, the estimated motion vectors have to be transmitted as additional side information. Recent research revealed that the data rate can be reduced by performing an additional motion estimation at the decoder. As only already decoded data is used, no additional data has to be transmitted. This paper addresses an improved hierarchical motion estimation algorithm to be used in a decoder-side motion estimation system. A special motion vector latching is used to be more robust for very small block sizes and to better adapt to object borders. With this technique, a dense motion vector field is estimated which reduces the rate by 6.9% in average compared to H.264 / AVC at the same quality.
当前的视频编码标准在编码器上执行运动估计,以便在编码之前预测帧。由于解码器不具有源帧,估计的运动矢量必须作为附加的侧信息传输。最近的研究表明,可以通过在解码器执行额外的运动估计来降低数据速率。因为只使用已经解码的数据,所以不需要传输额外的数据。本文提出了一种用于解码器侧运动估计系统的改进的分层运动估计算法。一种特殊的运动矢量锁存用于非常小的块大小,并更好地适应对象边界。使用该技术,估计了密集的运动矢量场,与相同质量的H.264 / AVC相比,平均降低了6.9%的速率。
{"title":"Decoder-side hierarchical motion estimation for dense vector fields","authors":"S. Klomp, Marco Munderloh, J. Ostermann","doi":"10.1109/PCS.2010.5702509","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702509","url":null,"abstract":"Current video coding standards perform motion estimation at the encoder to predict frames prior to coding them. Since the decoder does not possess the source frames, the estimated motion vectors have to be transmitted as additional side information. Recent research revealed that the data rate can be reduced by performing an additional motion estimation at the decoder. As only already decoded data is used, no additional data has to be transmitted. This paper addresses an improved hierarchical motion estimation algorithm to be used in a decoder-side motion estimation system. A special motion vector latching is used to be more robust for very small block sizes and to better adapt to object borders. With this technique, a dense motion vector field is estimated which reduces the rate by 6.9% in average compared to H.264 / AVC at the same quality.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123385004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
An improved low delay inter frame coding using template matching averaging 一种改进的基于模板匹配平均的低延迟帧间编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702511
Yoshinori Suzuki, C. Boon
This paper presents an efficient forward inter prediction method for video coding, targeting at low delay applications. The method applies the idea of template matching averaging (TMA) to the conventional motion compensated prediction (MCP). TMA forms the final predictor of a target block by averaging multiple numbers of candidates. While one of the candidate is specified by a motion vector, the remaining candidates are obtained based on the minimal matching error of a group of reconstructed pixels surrounding the target block, i.e., the template, against the reference frames. In this manner, additional predictors can be obtained without using explicit motion vector. In addition, the averaging of multiple predictors reduces coding noise residing in each of the predictors and hence contributes to improving the prediction efficiency. Simulation results show that the proposed scheme improves coding efficiency up to 4.5%, over the conventional MCP, without incurring coding delay due to backward prediction.
针对低时延应用,提出了一种高效的视频编码前向互预测方法。该方法将模板匹配平均(TMA)的思想应用于传统的运动补偿预测(MCP)。TMA通过对多个候选数据进行平均,形成目标块的最终预测器。当其中一个候选点由运动向量指定时,剩余的候选点是基于目标块(即模板)周围的一组重构像素对参考帧的最小匹配误差获得的。通过这种方式,可以在不使用显式运动向量的情况下获得额外的预测因子。此外,多个预测器的平均减少了存在于每个预测器中的编码噪声,从而有助于提高预测效率。仿真结果表明,与传统的MCP相比,该方案的编码效率提高了4.5%,且不会因后向预测而产生编码延迟。
{"title":"An improved low delay inter frame coding using template matching averaging","authors":"Yoshinori Suzuki, C. Boon","doi":"10.1109/PCS.2010.5702511","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702511","url":null,"abstract":"This paper presents an efficient forward inter prediction method for video coding, targeting at low delay applications. The method applies the idea of template matching averaging (TMA) to the conventional motion compensated prediction (MCP). TMA forms the final predictor of a target block by averaging multiple numbers of candidates. While one of the candidate is specified by a motion vector, the remaining candidates are obtained based on the minimal matching error of a group of reconstructed pixels surrounding the target block, i.e., the template, against the reference frames. In this manner, additional predictors can be obtained without using explicit motion vector. In addition, the averaging of multiple predictors reduces coding noise residing in each of the predictors and hence contributes to improving the prediction efficiency. Simulation results show that the proposed scheme improves coding efficiency up to 4.5%, over the conventional MCP, without incurring coding delay due to backward prediction.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"105 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121250360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Hyperspectral image compression suitable for spectral analysis application 高光谱图像压缩适用于光谱分析应用
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702582
Kazuma Shinoda, Y. Kosugi, Y. Murakami, Masahiro Yamaguchi, N. Ohyama
In the compression technique of hyperspectral image (HSI), PSNR of the reconstructed image is usually used for evaluating the performance of the coding results. For the spectral analysis applications of HSI, it is also important to consider the error in the result of spectral analysis. In the vegetation analysis, for example, the distortion of the vegetation index should be considered in addition to the distortion in the spectral data. This paper presents a HSI compression considering the error of both vegetation index and spectral data. The proposed method separates a hyperspectral data into spectral data for vegetation index and residual data. Both of the data are encoded by using a seamless coding individually. By holding the spectral channels required for vegetation index in the head of the code-stream, a precise vegetation analysis can be done in a low bit rate. Additionally, by decoding the residual data, the spectral data can be reconstructed in low distortion.
在高光谱图像的压缩技术中,通常使用重构图像的PSNR来评价编码结果的性能。对于恒生指数的光谱分析应用,考虑光谱分析结果中的误差也很重要。例如,在植被分析中,除了考虑光谱数据的畸变外,还应考虑植被指数的畸变。本文提出了一种考虑植被指数和光谱数据误差的HSI压缩方法。该方法将高光谱数据分离为植被指数光谱数据和残差数据。这两个数据分别使用无缝编码进行编码。通过在码流头部保留植被指数所需的频谱通道,可以在低比特率下进行精确的植被分析。此外,通过对残差数据进行解码,可以在低失真的情况下重建光谱数据。
{"title":"Hyperspectral image compression suitable for spectral analysis application","authors":"Kazuma Shinoda, Y. Kosugi, Y. Murakami, Masahiro Yamaguchi, N. Ohyama","doi":"10.1109/PCS.2010.5702582","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702582","url":null,"abstract":"In the compression technique of hyperspectral image (HSI), PSNR of the reconstructed image is usually used for evaluating the performance of the coding results. For the spectral analysis applications of HSI, it is also important to consider the error in the result of spectral analysis. In the vegetation analysis, for example, the distortion of the vegetation index should be considered in addition to the distortion in the spectral data. This paper presents a HSI compression considering the error of both vegetation index and spectral data. The proposed method separates a hyperspectral data into spectral data for vegetation index and residual data. Both of the data are encoded by using a seamless coding individually. By holding the spectral channels required for vegetation index in the head of the code-stream, a precise vegetation analysis can be done in a low bit rate. Additionally, by decoding the residual data, the spectral data can be reconstructed in low distortion.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"162 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116910460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Real-time Free Viewpoint Television for embedded systems 嵌入式系统实时免费视点电视
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702504
D. Aliprandi, E. Piccinelli
In this paper we describe an image-based rendering pipeline for interactive real-time Free Viewpoint Television (FTV) on embedded systems. Description of the processing steps and optimizations implemented targeting the hardware acceleration of a commercial programmable Graphics Processing Unit (GPU) is given. As a result, real-time view synthesis at 70 fps in XGA resolution has bee achieved. Restrictions and modifications introduced to support the application on OpenGL ES 2.0 based GPUs for embedded systems have also been discussed.
本文描述了一种用于嵌入式系统交互式实时自由视点电视(FTV)的基于图像的渲染管道。描述了针对商用可编程图形处理单元(GPU)硬件加速的处理步骤和实现的优化。因此,在XGA分辨率下实现了70 fps的实时视图合成。本文还讨论了为支持基于OpenGL ES 2.0的嵌入式系统gpu的应用而引入的限制和修改。
{"title":"Real-time Free Viewpoint Television for embedded systems","authors":"D. Aliprandi, E. Piccinelli","doi":"10.1109/PCS.2010.5702504","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702504","url":null,"abstract":"In this paper we describe an image-based rendering pipeline for interactive real-time Free Viewpoint Television (FTV) on embedded systems. Description of the processing steps and optimizations implemented targeting the hardware acceleration of a commercial programmable Graphics Processing Unit (GPU) is given. As a result, real-time view synthesis at 70 fps in XGA resolution has bee achieved. Restrictions and modifications introduced to support the application on OpenGL ES 2.0 based GPUs for embedded systems have also been discussed.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114271705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Video encoding with the original picture as the reference picture 视频编码以原始图片作为参考图片
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702520
Taiga Muromoto, N. Sagara, K. Sugiyama
Inter-picture prediction uses a local decoded picture for the reference, in order to avoid a mismatch between encoding and decoding. However, this scheme does not necessarily result in optimal coding efficiency since it requires encoding the processing altogether. Therefore, we study the use of the original picture as the reference. In this case, although the mismatch causes degradation of the picture quality, the bit amount is reduced. If optimal encoding is used at each macroblock, the overall performance may be improved. Therefore, we propose an adaptive method based on rate distortion optimization. The original picture is used only in the macroblock, if it is lower cost than the local decoded picture is used. Experimental results show a 0.1 to 1.0 dB gain in PSNR in each sequence. The adaptive method is shown to work successfully and the coding performance is improved without side effects.
图像间预测使用本地解码图像作为参考,以避免编码和解码之间的不匹配。然而,这种方案并不一定会产生最佳的编码效率,因为它需要对整个处理进行编码。因此,我们研究使用原始图片作为参考。在这种情况下,尽管不匹配导致图像质量下降,但比特量减少了。如果在每个宏块上使用最优编码,则整体性能可能会得到改善。因此,我们提出了一种基于速率失真优化的自适应方法。原始图片只在宏块中使用,如果它比本地解码图片的成本低。实验结果表明,每个序列的PSNR增益为0.1 ~ 1.0 dB。结果表明,自适应方法是有效的,并且在无副作用的情况下提高了编码性能。
{"title":"Video encoding with the original picture as the reference picture","authors":"Taiga Muromoto, N. Sagara, K. Sugiyama","doi":"10.1109/PCS.2010.5702520","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702520","url":null,"abstract":"Inter-picture prediction uses a local decoded picture for the reference, in order to avoid a mismatch between encoding and decoding. However, this scheme does not necessarily result in optimal coding efficiency since it requires encoding the processing altogether. Therefore, we study the use of the original picture as the reference. In this case, although the mismatch causes degradation of the picture quality, the bit amount is reduced. If optimal encoding is used at each macroblock, the overall performance may be improved. Therefore, we propose an adaptive method based on rate distortion optimization. The original picture is used only in the macroblock, if it is lower cost than the local decoded picture is used. Experimental results show a 0.1 to 1.0 dB gain in PSNR in each sequence. The adaptive method is shown to work successfully and the coding performance is improved without side effects.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122304983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D television system based on Integral Photography 基于积分摄影的三维电视系统
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702465
T. Mishina
Integral Photography (IP) is a photographic technique in which a lens array consisting of a large number of tiny lenses is used to capture and display three-dimensional (3D) images [1]. The displayed 3D images are optical real images, which give a natural 3D feeling in principle, without special viewing glasses. This is considered to be a suitable 3D display method for future 3D television systems. This paper describes an integral 3D television system in which Super Hi-Vision [2] is applied to IP.
集成摄影(integrated Photography, IP)是一种利用由大量微小透镜组成的透镜阵列来捕捉和显示三维图像的摄影技术。所显示的3D图像是光学真实图像,原则上给人一种自然的3D感觉,不需要特殊的观看眼镜。这被认为是一种适合未来3D电视系统的3D显示方法。本文介绍了一种将超高清[2]应用于IP的整体三维电视系统。
{"title":"3D television system based on Integral Photography","authors":"T. Mishina","doi":"10.1109/PCS.2010.5702465","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702465","url":null,"abstract":"Integral Photography (IP) is a photographic technique in which a lens array consisting of a large number of tiny lenses is used to capture and display three-dimensional (3D) images [1]. The displayed 3D images are optical real images, which give a natural 3D feeling in principle, without special viewing glasses. This is considered to be a suitable 3D display method for future 3D television systems. This paper describes an integral 3D television system in which Super Hi-Vision [2] is applied to IP.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129890270","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
A high efficiency coding framework for multiple image compression of circular camera array 一种高效的圆形相机阵列多图像压缩编码框架
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702505
Dongming Xue, Akira Kubota, Y. Hatori
Many existing multi-view video coding techniques remove inter-viewpoint redundancy by applying disparity compensation in conventional video coding frameworks, e.g., H264/MPEG4. However, conventional methodology works ineffectively as they ignore the special features of the inter-view-point disparity. This paper proposes a framework using virtual plane (VP) [1] for multi-view image compression, such that we can largely reduce the disparity compensation cost. Based on this VP predictor, we design a poxel[2] (probabilistic voxelized volume) framework, which integrates the information of the cameras in different view-points in the polar axis to obtain a more effective compression performance. In addition, considering the replay convenience of the multi-view video at the receiving side, we reform overhead information in polar axis at the sending side in advance.
许多现有的多视点视频编码技术通过在传统的视频编码框架(如H264/MPEG4)中应用视差补偿来消除视点间冗余。然而,传统的方法是无效的,因为它们忽略了访谈视点差异的特点。本文提出了一种利用虚拟平面(VP)[1]进行多视点图像压缩的框架,可以大大降低视差补偿成本。基于此VP预测器,我们设计了一个poxel[2](概率体素化体积)框架,该框架集成了极轴上不同视点的摄像机信息,以获得更有效的压缩性能。此外,考虑到接收端多视点视频的重放便利性,我们提前对发送端极轴上的开销信息进行了改造。
{"title":"A high efficiency coding framework for multiple image compression of circular camera array","authors":"Dongming Xue, Akira Kubota, Y. Hatori","doi":"10.1109/PCS.2010.5702505","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702505","url":null,"abstract":"Many existing multi-view video coding techniques remove inter-viewpoint redundancy by applying disparity compensation in conventional video coding frameworks, e.g., H264/MPEG4. However, conventional methodology works ineffectively as they ignore the special features of the inter-view-point disparity. This paper proposes a framework using virtual plane (VP) [1] for multi-view image compression, such that we can largely reduce the disparity compensation cost. Based on this VP predictor, we design a poxel[2] (probabilistic voxelized volume) framework, which integrates the information of the cameras in different view-points in the polar axis to obtain a more effective compression performance. In addition, considering the replay convenience of the multi-view video at the receiving side, we reform overhead information in polar axis at the sending side in advance.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126799005","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Complementary coding mode design based on R-D cost minimization for extending H.264 coding technology 扩展H.264编码技术的基于研发成本最小化的互补编码模式设计
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702549
T. Yoshino, S. Naito, S. Sakazawa, S. Matsumoto
To improve high resolution video coding efficiency under low bit-rate condition, an appropriate coding mode is required from an R-D optimization (RDO) perspective, although a coding mode defined within the H.264 standard is not always optimal for RDO criteria. With this in mind, we previously proposed extended SKIP modes with close-to-optimal R-D characteristics. However, the additional modes did not always satisfy the optimal R-D characteristics, especially for low bit-rate coding. In this paper, we propose an enhanced coding mode capable of providing a candidate corresponding to the minimum R-D cost by controlling the residual signal associated with the extended SKIP mode. The experimental result showed that the PSNR improvement against H.264 and our previous approach reached 0.42 dB and 0.24 dB in the maximum case, respectively.
为了提高低比特率条件下的高分辨率视频编码效率,从R-D优化(RDO)的角度来看,需要一种合适的编码模式,尽管H.264标准中定义的编码模式对于RDO标准并不总是最优的。考虑到这一点,我们之前提出了具有接近最优R-D特性的扩展SKIP模式。然而,附加模式并不总是满足最佳的R-D特性,特别是对于低比特率编码。在本文中,我们提出了一种增强的编码模式,能够通过控制与扩展SKIP模式相关的剩余信号来提供最小R-D成本对应的候选信号。实验结果表明,相对于H.264和我们之前的方法,PSNR的最大改进分别达到0.42 dB和0.24 dB。
{"title":"Complementary coding mode design based on R-D cost minimization for extending H.264 coding technology","authors":"T. Yoshino, S. Naito, S. Sakazawa, S. Matsumoto","doi":"10.1109/PCS.2010.5702549","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702549","url":null,"abstract":"To improve high resolution video coding efficiency under low bit-rate condition, an appropriate coding mode is required from an R-D optimization (RDO) perspective, although a coding mode defined within the H.264 standard is not always optimal for RDO criteria. With this in mind, we previously proposed extended SKIP modes with close-to-optimal R-D characteristics. However, the additional modes did not always satisfy the optimal R-D characteristics, especially for low bit-rate coding. In this paper, we propose an enhanced coding mode capable of providing a candidate corresponding to the minimum R-D cost by controlling the residual signal associated with the extended SKIP mode. The experimental result showed that the PSNR improvement against H.264 and our previous approach reached 0.42 dB and 0.24 dB in the maximum case, respectively.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"2011 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114469198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Diffusion filtering of depth maps in stereo video coding 立体视频编码中深度图的扩散滤波
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702493
G. Tech, K. Müller, T. Wiegand
A method for removing irrelevant information from depth maps in Video plus Depth coding is presented. The depth map is filtered in several iterations using a diffusional approach. In each iteration smoothing is carried out in local sample neighborhoods considering the distortion introduced to a rendered view. Smoothing is only applied when the rendered view is not affected. Therefore irrelevant edges and features in the depth map can be damped while the quality of the rendered view is retained. The processed depth maps can be coded at a reduced rate compared to unaltered data. Coding experiments show gains up to 0.5dB for the rendered view at the same bit rate.
提出了一种去除视频加深度编码中深度图中不相关信息的方法。深度图在几次迭代中使用扩散方法进行过滤。在每次迭代中,考虑到渲染视图中引入的畸变,对局部样本邻域进行平滑。平滑只在渲染视图不受影响时应用。因此,深度图中不相关的边缘和特征可以在保持渲染视图质量的同时被抑制。与未改变的数据相比,处理后的深度图可以以较低的速率进行编码。编码实验表明,在相同的比特率下,渲染视图的增益高达0.5dB。
{"title":"Diffusion filtering of depth maps in stereo video coding","authors":"G. Tech, K. Müller, T. Wiegand","doi":"10.1109/PCS.2010.5702493","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702493","url":null,"abstract":"A method for removing irrelevant information from depth maps in Video plus Depth coding is presented. The depth map is filtered in several iterations using a diffusional approach. In each iteration smoothing is carried out in local sample neighborhoods considering the distortion introduced to a rendered view. Smoothing is only applied when the rendered view is not affected. Therefore irrelevant edges and features in the depth map can be damped while the quality of the rendered view is retained. The processed depth maps can be coded at a reduced rate compared to unaltered data. Coding experiments show gains up to 0.5dB for the rendered view at the same bit rate.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129433182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
28th Picture Coding Symposium
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1