首页 > 最新文献

2011 IEEE 13th International Workshop on Multimedia Signal Processing最新文献

英文 中文
In-loop noise shaping based on pseudo noise injection and Wiener filtering 基于伪噪声注入和维纳滤波的环内噪声整形
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093804
K. Chono, Y. Senda
This paper proposes an in-loop noise shaping method based on the combination of pseudo noise injection and Wiener filtering for High Efficiency Video Coding (HEVC). In deblocking process, the proposed method injects pseudo noise into the vicinities of deblocked edges where signal-dependent coding noises are supposed to appear. The pseudo noise injection introduces a masking effect of signal-dependent noise with signal-independent noise. Since subsequently applied Wiener filtering accomplishes optimal noise reduction in a minimum-mean-squared-errors sense, it minimizes the deleterious impact of the pseudo noise injection on coding performance. Simulation results using HEVC Test Model software show that the proposed method successfully suppresses banding artifacts with a negligible impact on PSNR values, bit rates, and encoder/decoder runtime measures.
提出了一种基于伪噪声注入和维纳滤波相结合的高效视频编码(HEVC)环内噪声整形方法。在去块过程中,该方法在去块边缘的附近注入伪噪声,该区域应该出现与信号相关的编码噪声。伪噪声注入引入了信号相关噪声与信号无关噪声的掩蔽效应。由于随后应用维纳滤波在最小均方误差意义上实现了最佳降噪,因此它最大限度地减少了伪噪声注入对编码性能的有害影响。使用HEVC测试模型软件的仿真结果表明,该方法成功地抑制了带伪影,对PSNR值、比特率和编码器/解码器运行时间的影响可以忽略不计。
{"title":"In-loop noise shaping based on pseudo noise injection and Wiener filtering","authors":"K. Chono, Y. Senda","doi":"10.1109/MMSP.2011.6093804","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093804","url":null,"abstract":"This paper proposes an in-loop noise shaping method based on the combination of pseudo noise injection and Wiener filtering for High Efficiency Video Coding (HEVC). In deblocking process, the proposed method injects pseudo noise into the vicinities of deblocked edges where signal-dependent coding noises are supposed to appear. The pseudo noise injection introduces a masking effect of signal-dependent noise with signal-independent noise. Since subsequently applied Wiener filtering accomplishes optimal noise reduction in a minimum-mean-squared-errors sense, it minimizes the deleterious impact of the pseudo noise injection on coding performance. Simulation results using HEVC Test Model software show that the proposed method successfully suppresses banding artifacts with a negligible impact on PSNR values, bit rates, and encoder/decoder runtime measures.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114824989","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
ROI based video streaming for 3D remote rendering 基于ROI的3D远程渲染视频流
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093837
N. Tizon, Christina Moreno, M. Preda
This paper proposes a low computational method to perform ROI (Region Of Interest) based video encoding and adaptive streaming for remote rendering applications. The main objective of the proposed solution is to minimize the latency in the interactive loop even when facing poor transmission conditions. In order to do that, the knowledge of the depth map information provided by the rendering engine is exploited by the real-time video encoder to adapt the bitrate of the transmitted stream. Especially, thanks to an efficient coupling between the rendering and the video encoding stages, the macroblocks of each video frame are encoded with different quantization steps that follow an ROI partitioning. The details of this partitioning algorithm are provided as well with some implementation considerations. The simulation results demonstrate the benefit of our adaptive approach from the user experience point of view.
本文提出了一种基于感兴趣区域的视频编码和自适应流的低计算方法。所提出的解决方案的主要目标是即使面对较差的传输条件,也要最小化交互环路中的延迟。为了做到这一点,实时视频编码器利用渲染引擎提供的深度图信息来调整传输流的比特率。特别是,由于渲染和视频编码阶段之间的有效耦合,每个视频帧的宏块按照ROI划分用不同的量化步骤进行编码。本文还提供了这种分区算法的细节,以及一些实现注意事项。仿真结果从用户体验的角度证明了自适应方法的优点。
{"title":"ROI based video streaming for 3D remote rendering","authors":"N. Tizon, Christina Moreno, M. Preda","doi":"10.1109/MMSP.2011.6093837","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093837","url":null,"abstract":"This paper proposes a low computational method to perform ROI (Region Of Interest) based video encoding and adaptive streaming for remote rendering applications. The main objective of the proposed solution is to minimize the latency in the interactive loop even when facing poor transmission conditions. In order to do that, the knowledge of the depth map information provided by the rendering engine is exploited by the real-time video encoder to adapt the bitrate of the transmitted stream. Especially, thanks to an efficient coupling between the rendering and the video encoding stages, the macroblocks of each video frame are encoded with different quantization steps that follow an ROI partitioning. The details of this partitioning algorithm are provided as well with some implementation considerations. The simulation results demonstrate the benefit of our adaptive approach from the user experience point of view.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132233033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
A fast video stabilization algorithm based on block matching and edge completion 基于块匹配和边缘补全的快速视频稳像算法
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093781
Chongwu Tang, Xiaokang Yang, Li Chen, Guangtao Zhai
Purpose of video stabilization is to register the frames of a video sequence with relative motions between each other to yield a stable video of higher perceptual quality. In this paper we focus on the problem of fast and robust video stabilization for the same scene based on temporal block matching. We use VoD principle to find the local motion vectors between adjacent frames, and then use statistical analysis to generate the global vibrant motion vector. After motion compensation, we further design an edge completion algorithm incorporating mosaicking and inpainting of neighbour frames, so as to reduce the impact of error propagation. Experimental results and comparative studies will be provided to justify the effectiveness of the proposed algorithm.
视频防抖的目的是对具有相对运动的视频序列的帧进行注册,以产生具有较高感知质量的稳定视频。本文主要研究基于时序块匹配的同一场景快速鲁棒视频稳像问题。首先利用点播原理找到相邻帧之间的局部运动矢量,然后利用统计分析方法生成全局动态运动矢量。在运动补偿之后,我们进一步设计了一种结合相邻帧的拼接和补全的边缘补全算法,以减少误差传播的影响。将提供实验结果和比较研究来证明所提出算法的有效性。
{"title":"A fast video stabilization algorithm based on block matching and edge completion","authors":"Chongwu Tang, Xiaokang Yang, Li Chen, Guangtao Zhai","doi":"10.1109/MMSP.2011.6093781","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093781","url":null,"abstract":"Purpose of video stabilization is to register the frames of a video sequence with relative motions between each other to yield a stable video of higher perceptual quality. In this paper we focus on the problem of fast and robust video stabilization for the same scene based on temporal block matching. We use VoD principle to find the local motion vectors between adjacent frames, and then use statistical analysis to generate the global vibrant motion vector. After motion compensation, we further design an edge completion algorithm incorporating mosaicking and inpainting of neighbour frames, so as to reduce the impact of error propagation. Experimental results and comparative studies will be provided to justify the effectiveness of the proposed algorithm.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129312506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Motion-adaptive quantization and reconstruction technique for distributed video coding 分布式视频编码的运动自适应量化与重构技术
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093782
Aniruddha Shirahatti, Joohee Kim
Distributed video coding is a new paradigm based on two information theoretical results by Slepian-Wolf and Wyner-Ziv. The architectures designed so far have invariably made use of the uniform scalar quantization schemes along with a few attempts to make the schemes more adaptive. Quantization is one of the major contributors to the large performance gap between conventional video coding standards and distributed video coding. In this paper, an attempt is made to improve the performance of the Wyner-Ziv video coding by making the quantization algorithm more adaptive to the motion content of the video sequence without significantly increasing the encoder complexity. The proposed method also exploits the temporal correlation to provide for online correlation noise classification. Hence, the improved reconstruction technique which uses the correlation noise information is more adaptive to the motion content. Simulation results show that the proposed motion-adaptive quantization and reconstruction technique achieves improved rate-distortion performance.
分布式视频编码是基于Slepian-Wolf和Wyner-Ziv两个信息理论成果的一种新的编码范式。迄今为止设计的体系结构总是使用统一的标量量化方案以及一些使方案更具适应性的尝试。量化是造成传统视频编码标准与分布式视频编码之间存在较大性能差距的主要原因之一。本文试图在不显著增加编码器复杂度的前提下,使量化算法更能适应视频序列的运动内容,从而提高Wyner-Ziv视频编码的性能。该方法还利用时间相关性提供在线相关噪声分类。因此,利用相关噪声信息的改进重建技术对运动内容的适应性更强。仿真结果表明,所提出的运动自适应量化和重建技术取得了较好的率失真性能。
{"title":"Motion-adaptive quantization and reconstruction technique for distributed video coding","authors":"Aniruddha Shirahatti, Joohee Kim","doi":"10.1109/MMSP.2011.6093782","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093782","url":null,"abstract":"Distributed video coding is a new paradigm based on two information theoretical results by Slepian-Wolf and Wyner-Ziv. The architectures designed so far have invariably made use of the uniform scalar quantization schemes along with a few attempts to make the schemes more adaptive. Quantization is one of the major contributors to the large performance gap between conventional video coding standards and distributed video coding. In this paper, an attempt is made to improve the performance of the Wyner-Ziv video coding by making the quantization algorithm more adaptive to the motion content of the video sequence without significantly increasing the encoder complexity. The proposed method also exploits the temporal correlation to provide for online correlation noise classification. Hence, the improved reconstruction technique which uses the correlation noise information is more adaptive to the motion content. Simulation results show that the proposed motion-adaptive quantization and reconstruction technique achieves improved rate-distortion performance.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"125 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126872220","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Saliency-based visualization for image search 基于显著性的图像搜索可视化
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093833
Jiajie Hu, Bin Jin, Weiyao Lin, Jun Huang, Hangzai Luo, Zhenzhong Chen, Hongxiang Li
In this paper, we propose a novel algorithm for improving and visualizing image search results. The proposed algorithm improves user's image search experience by three steps: (1) re-rank the initial image search results by the random walk refinement based on visual consistency and saliency cues, (2) project the re-ranked images into a 2-dimentional panel according to their saliency information and correlations, (3) detect and extract the saliency regions in each image for final visualization. To evaluate the performance of our algorithm, user study has been conducted. Experimental results demonstrate that our visualization algorithm provides more pleasing image search experience than the conventional image search methods.
在本文中,我们提出了一种改进和可视化图像搜索结果的新算法。该算法通过三个步骤提高用户的图像搜索体验:(1)通过基于视觉一致性和显著性线索的随机行走细化对初始图像搜索结果进行重新排序;(2)根据图像的显著性信息和相关性将重新排序的图像投影到二维面板中;(3)检测并提取每个图像中的显著性区域以实现最终的可视化。为了评估我们算法的性能,我们进行了用户研究。实验结果表明,该可视化算法比传统的图像搜索方法提供了更令人满意的图像搜索体验。
{"title":"Saliency-based visualization for image search","authors":"Jiajie Hu, Bin Jin, Weiyao Lin, Jun Huang, Hangzai Luo, Zhenzhong Chen, Hongxiang Li","doi":"10.1109/MMSP.2011.6093833","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093833","url":null,"abstract":"In this paper, we propose a novel algorithm for improving and visualizing image search results. The proposed algorithm improves user's image search experience by three steps: (1) re-rank the initial image search results by the random walk refinement based on visual consistency and saliency cues, (2) project the re-ranked images into a 2-dimentional panel according to their saliency information and correlations, (3) detect and extract the saliency regions in each image for final visualization. To evaluate the performance of our algorithm, user study has been conducted. Experimental results demonstrate that our visualization algorithm provides more pleasing image search experience than the conventional image search methods.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122201342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-hypothesis transform domain Wyner-Ziv video coding including optical flow 包含光流的多假设变换域Wyner-Ziv视频编码
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093771
Xin Huang, L. L. Rakêt, Huynh Van Luong, M. Nielsen, F. Lauze, Søren Forchhammer
Transform Domain Wyner-Ziv (TDWZ) video coding is an efficient Distributed Video coding solution providing new features such as low complexity encoding, by mainly exploiting the source statistics at the decoder based on the availability of decoder side information. The accuracy of the decoder side information has a major impact on the performance of TDWZ. In this paper, a novel multi-hypothesis based TDWZ video coding is presented to exploit the redundancy between multiple side information and the source information. The decoder used optical flow for side information calculation. Compared with the best available single estimation mode TDWZ, the proposed multi-hypothesis based TDWZ achieves robustly better Rate-Distortion (RD) performance and the overall improvement is up to 0.6 dB at high bitrate and up to 2 dB compared with the DISCOVER TDWZ video codec.
变换域Wyner-Ziv (TDWZ)视频编码是一种高效的分布式视频编码解决方案,通过基于解码器侧信息的可用性,主要利用解码器的源统计信息,提供了诸如低复杂度编码等新特性。解码器侧信息的准确性对TDWZ的性能有很大的影响。本文提出了一种新的基于多假设的TDWZ视频编码方法,利用了多侧信息与源信息之间的冗余性。解码器采用光流进行侧信息计算。与现有的最佳单估计模式TDWZ相比,所提出的基于多假设的TDWZ具有更好的鲁棒率失真(RD)性能,在高比特率下总体改进可达0.6 dB,与DISCOVER TDWZ视频编解码器相比可达2 dB。
{"title":"Multi-hypothesis transform domain Wyner-Ziv video coding including optical flow","authors":"Xin Huang, L. L. Rakêt, Huynh Van Luong, M. Nielsen, F. Lauze, Søren Forchhammer","doi":"10.1109/MMSP.2011.6093771","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093771","url":null,"abstract":"Transform Domain Wyner-Ziv (TDWZ) video coding is an efficient Distributed Video coding solution providing new features such as low complexity encoding, by mainly exploiting the source statistics at the decoder based on the availability of decoder side information. The accuracy of the decoder side information has a major impact on the performance of TDWZ. In this paper, a novel multi-hypothesis based TDWZ video coding is presented to exploit the redundancy between multiple side information and the source information. The decoder used optical flow for side information calculation. Compared with the best available single estimation mode TDWZ, the proposed multi-hypothesis based TDWZ achieves robustly better Rate-Distortion (RD) performance and the overall improvement is up to 0.6 dB at high bitrate and up to 2 dB compared with the DISCOVER TDWZ video codec.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124550261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Directional samples reordering for intra residual transform 残差变换的方向样本重排序
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093814
Q. Wu, Hongliang Li, Tiantang Chen
In this paper, a directional samples reordering (DSR) based algorithm was proposed for intra residual data transform. To make the residual more suitable for discrete cosine transform, the diagonal edge in arbitrary size of intra prediction block can be rotated to a regular horizontal edge by reordering the samples in the block. The intra residual data will be used to implement 2D discrete cosine transform after DSR procedure. Experimental results show that up to 0.5967 dB and on average 0.4372 dB gain can be achieved for CIF sequence in high bitrate with the proposed algorithm at high complexity mode than H.264 intra coding.
提出了一种基于方向性样本重排序(DSR)的残差数据变换算法。为了使残差更适合于离散余弦变换,可以通过对块内样本重新排序,将任意大小的预测块内的对角线边缘旋转为规则的水平边缘。在DSR处理后,利用残差数据进行二维离散余弦变换。实验结果表明,与H.264帧内编码相比,该算法在高码率下的CIF序列增益可达0.5967 dB,平均增益为0.4372 dB。
{"title":"Directional samples reordering for intra residual transform","authors":"Q. Wu, Hongliang Li, Tiantang Chen","doi":"10.1109/MMSP.2011.6093814","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093814","url":null,"abstract":"In this paper, a directional samples reordering (DSR) based algorithm was proposed for intra residual data transform. To make the residual more suitable for discrete cosine transform, the diagonal edge in arbitrary size of intra prediction block can be rotated to a regular horizontal edge by reordering the samples in the block. The intra residual data will be used to implement 2D discrete cosine transform after DSR procedure. Experimental results show that up to 0.5967 dB and on average 0.4372 dB gain can be achieved for CIF sequence in high bitrate with the proposed algorithm at high complexity mode than H.264 intra coding.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121646374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sample adaptive offset for HEVC HEVC的样本自适应偏移
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093807
Chih-Ming Fu, Ching-Yeh Chen, Yu-Wen Huang, S. Lei
A new video coding tool, sample adaptive offset (SAO), is introduced in this paper. SAO has been adopted into the Working Draft of the new video coding standard, High-Efficiency Video Coding (HEVC). The SAO is located after deblocking in the video coding loop. The concept of SAO is to classify reconstructed pixels into different categories and then reduce the distortion by simply adding an offset for each category of pixels. The pixel intensity and edge properties are used for pixel classification. To further improve the coding efficiency, a picture can be divided into regions for localization of offset parameters. Simulation results show that SAO can achieve on average 2% bit rate reduction and up to 6% bit rate reduction. The run time increases for encoders and decoders are only 2%.
本文介绍了一种新的视频编码工具——采样自适应偏移(SAO)。SAO已被纳入新的视频编码标准——高效视频编码(HEVC)的工作草案中。SAO位于视频编码循环中的块化之后。SAO的概念是将重建的像素分类为不同的类别,然后通过简单地为每个类别的像素添加偏移量来减少畸变。像素强度和边缘属性用于像素分类。为了进一步提高编码效率,可以将图像划分为多个区域进行偏移量参数的定位。仿真结果表明,SAO可以实现平均2%的比特率降低,最高可达6%的比特率降低。编码器和解码器的运行时间仅增加2%。
{"title":"Sample adaptive offset for HEVC","authors":"Chih-Ming Fu, Ching-Yeh Chen, Yu-Wen Huang, S. Lei","doi":"10.1109/MMSP.2011.6093807","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093807","url":null,"abstract":"A new video coding tool, sample adaptive offset (SAO), is introduced in this paper. SAO has been adopted into the Working Draft of the new video coding standard, High-Efficiency Video Coding (HEVC). The SAO is located after deblocking in the video coding loop. The concept of SAO is to classify reconstructed pixels into different categories and then reduce the distortion by simply adding an offset for each category of pixels. The pixel intensity and edge properties are used for pixel classification. To further improve the coding efficiency, a picture can be divided into regions for localization of offset parameters. Simulation results show that SAO can achieve on average 2% bit rate reduction and up to 6% bit rate reduction. The run time increases for encoders and decoders are only 2%.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132596121","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 72
On residual quad-tree coding in HEVC HEVC中残差四叉树编码
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093805
Y. H. Tan, Chuohao Yeo, Hui Li Tan, Zhengguo Li
In the current working draft of HEVC, residual quad-tree (RQT) coding is used to encode prediction residuals in both Intra and Inter coding units (CU). However, the rationale for using RQT as a coding tool is different in the two cases. For Intra prediction units, RQT provides an efficient syntax for coding a number of sub-blocks with the same intra prediction mode. For Inter CUs, RQT adapts to the spatial-frequency variations of the CU, using as large a transform size as possible while catering to local variations in residual statistics. While providing coding gains, effective use of RQT currently requires an exhaustive search of all possible combinations of transform sizes within a block. In this paper, we exploit our insights to develop two fast RQT algorithms, each designed to meet the needs of Intra and Inter prediction residual coding.
在当前的HEVC工作草案中,残差四叉树(RQT)编码在Intra和Inter编码单元(CU)中对预测残差进行编码。然而,在这两种情况下,使用RQT作为编码工具的基本原理是不同的。对于内部预测单元,RQT提供了一种有效的语法,可以用相同的内部预测模式对许多子块进行编码。对于Inter CU, RQT适应CU的空间-频率变化,使用尽可能大的变换大小,同时满足残差统计量的局部变化。在提供编码增益的同时,RQT的有效使用目前需要对块内所有可能的转换大小组合进行详尽的搜索。在本文中,我们利用我们的见解开发了两种快速RQT算法,每种算法都旨在满足内部和内部预测残差编码的需求。
{"title":"On residual quad-tree coding in HEVC","authors":"Y. H. Tan, Chuohao Yeo, Hui Li Tan, Zhengguo Li","doi":"10.1109/MMSP.2011.6093805","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093805","url":null,"abstract":"In the current working draft of HEVC, residual quad-tree (RQT) coding is used to encode prediction residuals in both Intra and Inter coding units (CU). However, the rationale for using RQT as a coding tool is different in the two cases. For Intra prediction units, RQT provides an efficient syntax for coding a number of sub-blocks with the same intra prediction mode. For Inter CUs, RQT adapts to the spatial-frequency variations of the CU, using as large a transform size as possible while catering to local variations in residual statistics. While providing coding gains, effective use of RQT currently requires an exhaustive search of all possible combinations of transform sizes within a block. In this paper, we exploit our insights to develop two fast RQT algorithms, each designed to meet the needs of Intra and Inter prediction residual coding.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"94 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125245836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Selective pixel interpolation for spatial error concealment 空间误差隐藏的选择性像素插值
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093828
Yi Ge, Bo Yan, Kairan Sun, H. Gharavi
This paper proposes an effective algorithm for spatial error concealment with accurate edge detection and partitioning interpolation. Firstly, a new method is used for detecting possible edge pixels and their matching pixels around the lost block. Then, the true edge lines can be determined, with which the lost block is partitioned. Finally, based on the partition result, each lost pixel can be interpolated with correct reference pixels, which are in the same region with the lost pixel. Experimental results show that the proposed spatial error concealment method is obviously superior to the previous methods for different sequences by up to 4.04 dB.
提出了一种基于精确边缘检测和分区插值的有效空间错误隐藏算法。首先,采用一种新的方法检测丢失块周围可能的边缘像素及其匹配像素。然后,可以确定真正的边缘线,并用它来分割丢失的块。最后,根据分割结果,将每个丢失的像素用正确的参考像素进行插值,这些参考像素与丢失的像素在同一区域。实验结果表明,对于不同的序列,本文提出的空间误差隐藏方法明显优于先前的方法,误差最大可达4.04 dB。
{"title":"Selective pixel interpolation for spatial error concealment","authors":"Yi Ge, Bo Yan, Kairan Sun, H. Gharavi","doi":"10.1109/MMSP.2011.6093828","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093828","url":null,"abstract":"This paper proposes an effective algorithm for spatial error concealment with accurate edge detection and partitioning interpolation. Firstly, a new method is used for detecting possible edge pixels and their matching pixels around the lost block. Then, the true edge lines can be determined, with which the lost block is partitioned. Finally, based on the partition result, each lost pixel can be interpolated with correct reference pixels, which are in the same region with the lost pixel. Experimental results show that the proposed spatial error concealment method is obviously superior to the previous methods for different sequences by up to 4.04 dB.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126607403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2011 IEEE 13th International Workshop on Multimedia Signal Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1