首页 > 最新文献

28th Picture Coding Symposium最新文献

英文 中文
Vision field capturing and its applications in 3DTV 视野捕捉及其在3DTV中的应用
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702457
Qionghai Dai, Xiangyang Ji, Xun Cao
3D video capturing acquires the visual information in 3D manner, which possesses the first step of the entire 3DTV system chain before 3D coding, transmission and visualization. The 3D capturing plays an important role because precise 3D visual capturing will benefit the whole 3DTV system. During the past decades, various kinds of capturing system have been built for different applications such as FTV[1], 3DTV, 3D movie, etc. As the cost of sensors reduces in recent years, a lot of systems utilize multiple cameras to acquire visual information, which is called multiview capturing. 3D information can be further extracted through multiview geometry. We will first give a brief review of these multiview systems and analyze their relationship from the perspective of plenoptic function [2]. Along with the multiple cameras, a lot of systems also make use of multiple lights to control the illumination condition. A new concept of vision field is presented in this talk according to the view-light-time subspace, which can be derived from the plenoptic function. The features and applications for each capturing system will be emphasized as well as the important issues in capturing like synchronization and calibration. Besides the multiple camera systems, some new techniques using TOF (time-offlight) camera [3] and 3D scanner will also be included in this talk.
3D视频采集以3D的方式获取视觉信息,是整个3DTV系统链中3D编码、传输和可视化前的第一步。三维捕捉是整个三维电视系统的重要组成部分,精确的三维视觉捕捉对整个系统的发展至关重要。在过去的几十年里,针对不同的应用建立了各种捕获系统,如FTV[1]、3DTV、3D电影等。近年来,随着传感器成本的降低,许多系统采用多摄像头采集视觉信息,这被称为多视图捕获。通过多视图几何可以进一步提取三维信息。我们将首先简要回顾这些多视点系统,并从全视函数的角度分析它们之间的关系[2]。随着多摄像头的出现,很多系统也利用多灯来控制照明条件。本文提出了一种新的视野概念,即视光时子空间,它可以由全视函数导出。本文将重点介绍每一种捕获系统的特点和应用,以及捕获过程中的重要问题,如同步和校准。除了多相机系统外,一些使用TOF (time-offlight)相机[3]和3D扫描仪的新技术也将在本次演讲中介绍。
{"title":"Vision field capturing and its applications in 3DTV","authors":"Qionghai Dai, Xiangyang Ji, Xun Cao","doi":"10.1109/PCS.2010.5702457","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702457","url":null,"abstract":"3D video capturing acquires the visual information in 3D manner, which possesses the first step of the entire 3DTV system chain before 3D coding, transmission and visualization. The 3D capturing plays an important role because precise 3D visual capturing will benefit the whole 3DTV system. During the past decades, various kinds of capturing system have been built for different applications such as FTV[1], 3DTV, 3D movie, etc. As the cost of sensors reduces in recent years, a lot of systems utilize multiple cameras to acquire visual information, which is called multiview capturing. 3D information can be further extracted through multiview geometry. We will first give a brief review of these multiview systems and analyze their relationship from the perspective of plenoptic function [2]. Along with the multiple cameras, a lot of systems also make use of multiple lights to control the illumination condition. A new concept of vision field is presented in this talk according to the view-light-time subspace, which can be derived from the plenoptic function. The features and applications for each capturing system will be emphasized as well as the important issues in capturing like synchronization and calibration. Besides the multiple camera systems, some new techniques using TOF (time-offlight) camera [3] and 3D scanner will also be included in this talk.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"20 6 Pt 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130842882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
RBF-based VBR controller for real-time H.264/SVC video coding 基于rbf的实时H.264/SVC视频编码的VBR控制器
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702522
S. Rodríguez, F. Díaz-de-María
In this paper we propose a novel VBR controller for real-time H.264/SVC video coding. Since consecutive pictures within the same scene often exhibit similar degrees of complexity, the proposed VBR controller allows for just an incremental variation of QP with respect to that of the previous picture, so preventing unnecessary QP fluctuations. For this purpose, an RBF network has been carefully designed to estimate the QP increment at each dependency (spatial or CGS) layer. A mobile live streaming application scenario was simulated to assess the performance of the proposed VBR controller, which was compared to a recently proposed CBR controller for H.264/SVC. The experimental results show a remarkably consistent quality, notably outperforming the reference CBR controller.
本文提出了一种用于实时H.264/SVC视频编码的VBR控制器。由于同一场景中的连续图像通常表现出相似的复杂程度,因此所提出的VBR控制器只允许相对于前一个图像的QP增量变化,从而防止不必要的QP波动。为此,RBF网络被精心设计来估计每个依赖层(空间层或CGS层)的QP增量。通过对移动直播应用场景的仿真,评估了所提出的VBR控制器的性能,并与最近提出的基于H.264/SVC的CBR控制器进行了比较。实验结果表明,该控制器具有良好的一致性,明显优于参考CBR控制器。
{"title":"RBF-based VBR controller for real-time H.264/SVC video coding","authors":"S. Rodríguez, F. Díaz-de-María","doi":"10.1109/PCS.2010.5702522","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702522","url":null,"abstract":"In this paper we propose a novel VBR controller for real-time H.264/SVC video coding. Since consecutive pictures within the same scene often exhibit similar degrees of complexity, the proposed VBR controller allows for just an incremental variation of QP with respect to that of the previous picture, so preventing unnecessary QP fluctuations. For this purpose, an RBF network has been carefully designed to estimate the QP increment at each dependency (spatial or CGS) layer. A mobile live streaming application scenario was simulated to assess the performance of the proposed VBR controller, which was compared to a recently proposed CBR controller for H.264/SVC. The experimental results show a remarkably consistent quality, notably outperforming the reference CBR controller.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"157 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126900942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Generating subject oriented codec by evolutionary approach 用进化方法生成面向主题的编解码器
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702512
Masaaki Matsumura, Seishi Takamura, H. Jozawa
Many image/video codecs are constructed by the combination of various coding tools such as block division/scanning, branch selection and entropy coders. Codec researchers are developing new coding tools, and seeking versatile combinations that offer improved coding efficiency for various images/videos. However, because of the huge amount of the combination, deriving the best combination is impossible by man-power seeking. In this paper, we propose an automatic optimization method for deriving the combination that suits for categorized pictures. We prepare some categorised pictures, and optimize the combination for each category. In the case of optimization for lossless image coding, our method achieves a bit-rate reduction of over 2.8% (maximum) compared to the combination that offers the best bit-rate averagely prepared beforehand.
许多图像/视频编解码器是由各种编码工具如分块/扫描、分支选择和熵编码器组合而成的。编解码器研究人员正在开发新的编码工具,并寻求多种组合,以提高各种图像/视频的编码效率。然而,由于组合的数量巨大,单靠人力寻找是不可能得出最佳组合的。在本文中,我们提出了一种自动优化方法来推导适合分类图片的组合。我们准备了一些分类图片,并对每个分类的组合进行了优化。在优化无损图像编码的情况下,与提供最佳比特率的组合相比,我们的方法实现了超过2.8%(最大)的比特率降低。
{"title":"Generating subject oriented codec by evolutionary approach","authors":"Masaaki Matsumura, Seishi Takamura, H. Jozawa","doi":"10.1109/PCS.2010.5702512","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702512","url":null,"abstract":"Many image/video codecs are constructed by the combination of various coding tools such as block division/scanning, branch selection and entropy coders. Codec researchers are developing new coding tools, and seeking versatile combinations that offer improved coding efficiency for various images/videos. However, because of the huge amount of the combination, deriving the best combination is impossible by man-power seeking. In this paper, we propose an automatic optimization method for deriving the combination that suits for categorized pictures. We prepare some categorised pictures, and optimize the combination for each category. In the case of optimization for lossless image coding, our method achieves a bit-rate reduction of over 2.8% (maximum) compared to the combination that offers the best bit-rate averagely prepared beforehand.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121300368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A fast graph cut algorithm for disparity estimation 视差估计的快速图割算法
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702499
Cheng-Wei Chou, J. Tsai, H. Hang, Hung-Chih Lin
In this paper, we propose a fast graph cut (GC) algorithm for disparity estimation. Two accelerating techniques are suggested: one is the early termination rule, and the other is prioritizing the α-β swap pair search order. Our simulations show that the proposed fast GC algorithm outperforms the original GC scheme by 210% in the average computation time while its disparity estimation quality is almost similar to that of the original GC.
本文提出了一种用于视差估计的快速图割(GC)算法。提出了两种加速技术:一种是提前终止规则,另一种是优先考虑α-β交换对的搜索顺序。仿真结果表明,本文提出的快速GC算法在平均计算时间上比原GC方案提高了210%,而视差估计质量与原GC方案基本相当。
{"title":"A fast graph cut algorithm for disparity estimation","authors":"Cheng-Wei Chou, J. Tsai, H. Hang, Hung-Chih Lin","doi":"10.1109/PCS.2010.5702499","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702499","url":null,"abstract":"In this paper, we propose a fast graph cut (GC) algorithm for disparity estimation. Two accelerating techniques are suggested: one is the early termination rule, and the other is prioritizing the α-β swap pair search order. Our simulations show that the proposed fast GC algorithm outperforms the original GC scheme by 210% in the average computation time while its disparity estimation quality is almost similar to that of the original GC.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"86 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120895236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Fast and efficient Gaussian noise image restoration algorithm by spatially adaptive filtering 基于空间自适应滤波的快速高效高斯噪声图像恢复算法
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702438
Tuan-Anh Nguyen, M. Kim, Min-Cheol Hong
In this paper, we propose a spatially adaptive noise removal algorithm using local statistics that consists of two stages: noise detection and removal. To corporate desirable properties into denoising process, the local weighted mean, local weighted activity, and local maximum are defined. With these local statistics, the noise detection function is defined and a modified Gaussian filter is used to suppress the detected noise components. The experimental results demonstrate the effectiveness of the proposed algorithm.
在本文中,我们提出了一种基于局部统计的空间自适应噪声去除算法,该算法包括两个阶段:噪声检测和去除。定义了局部加权均值、局部加权活度和局部极大值,将期望的性质融入到去噪过程中。利用这些局部统计量,定义噪声检测函数,并使用改进的高斯滤波器抑制检测到的噪声分量。实验结果证明了该算法的有效性。
{"title":"Fast and efficient Gaussian noise image restoration algorithm by spatially adaptive filtering","authors":"Tuan-Anh Nguyen, M. Kim, Min-Cheol Hong","doi":"10.1109/PCS.2010.5702438","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702438","url":null,"abstract":"In this paper, we propose a spatially adaptive noise removal algorithm using local statistics that consists of two stages: noise detection and removal. To corporate desirable properties into denoising process, the local weighted mean, local weighted activity, and local maximum are defined. With these local statistics, the noise detection function is defined and a modified Gaussian filter is used to suppress the detected noise components. The experimental results demonstrate the effectiveness of the proposed algorithm.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"271 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115269714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
FFT-based full-search block matching using overlap-add method 基于fft的全搜索块匹配的重叠添加方法
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702570
H. Sasaki, Z. Li, H. Kiya
One category of fast full-search block matching algorithms (BMAs) is based on the fast Fourier transformation (FFT). In conventional methods in this category, the macroblock size must be adjusted to the search window size by zero-padding. In these methods, the memory consumption and computational complexity heavily depend on the size difference between the macroblock and the search window. Thus, we propose a novel FFT-based BMA to solve this problem. The proposed method divides the search window into multiple sub search windows to versatilely control the difference between the macroblock and the search window sizes. Simulation results show the effectiveness of the proposed method.
一类基于快速傅里叶变换的快速全搜索块匹配算法(BMAs)。在此类别的常规方法中,必须通过零填充将宏块大小调整为搜索窗口大小。在这些方法中,内存消耗和计算复杂度很大程度上取决于宏块和搜索窗口之间的大小差异。因此,我们提出了一种新的基于fft的BMA来解决这个问题。该方法将搜索窗口划分为多个子搜索窗口,以灵活地控制宏块与搜索窗口大小之间的差异。仿真结果表明了该方法的有效性。
{"title":"FFT-based full-search block matching using overlap-add method","authors":"H. Sasaki, Z. Li, H. Kiya","doi":"10.1109/PCS.2010.5702570","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702570","url":null,"abstract":"One category of fast full-search block matching algorithms (BMAs) is based on the fast Fourier transformation (FFT). In conventional methods in this category, the macroblock size must be adjusted to the search window size by zero-padding. In these methods, the memory consumption and computational complexity heavily depend on the size difference between the macroblock and the search window. Thus, we propose a novel FFT-based BMA to solve this problem. The proposed method divides the search window into multiple sub search windows to versatilely control the difference between the macroblock and the search window sizes. Simulation results show the effectiveness of the proposed method.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121023018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Distributed video coding based on adaptive slice size using received motion vectors 基于接收到的运动矢量自适应切片大小的分布式视频编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702481
Kyung-Yeon Min, Seanae Park, Donggyu Sim
In this paper, we propose a new distributed video coding (DVC) method based on adaptive slice size using received motion vectors (MVs). In the proposed algorithm, the MVs estimated at a DVC decoder are transmitted to a corresponding encoder. In the proposed encoder, a predicted side information (PSI) is reconstructed with the transmitted MVs and key frames. Therefore, the PSI can be generated same to side information (SI) at the decoder. We can, also, calculate an exact crossover probability between the SI and original input frame using PSI and the original frame. As a result, the proposed method can transmit minimum parity bits to maximize error correction ability of a channel decoder with minimal computational complexity. Experimental results show that the proposed algorithm is better than several conventional DVC methods.
在本文中,我们提出了一种基于自适应切片大小的分布式视频编码(DVC)方法。在该算法中,在DVC解码器估计的mv被传输到相应的编码器。在该编码器中,利用传输的mv和关键帧重构预测侧信息(PSI)。因此,PSI可以在解码器处生成同侧信息(SI)。我们还可以使用PSI和原始帧计算SI和原始输入帧之间的精确交叉概率。因此,该方法可以传输最小的奇偶校验位,以最小的计算复杂度最大化信道解码器的纠错能力。实验结果表明,该算法优于几种传统的DVC方法。
{"title":"Distributed video coding based on adaptive slice size using received motion vectors","authors":"Kyung-Yeon Min, Seanae Park, Donggyu Sim","doi":"10.1109/PCS.2010.5702481","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702481","url":null,"abstract":"In this paper, we propose a new distributed video coding (DVC) method based on adaptive slice size using received motion vectors (MVs). In the proposed algorithm, the MVs estimated at a DVC decoder are transmitted to a corresponding encoder. In the proposed encoder, a predicted side information (PSI) is reconstructed with the transmitted MVs and key frames. Therefore, the PSI can be generated same to side information (SI) at the decoder. We can, also, calculate an exact crossover probability between the SI and original input frame using PSI and the original frame. As a result, the proposed method can transmit minimum parity bits to maximize error correction ability of a channel decoder with minimal computational complexity. Experimental results show that the proposed algorithm is better than several conventional DVC methods.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122772308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
VQ based data hiding method for still images by tree-structured links 基于VQ的树状链接静态图像数据隐藏方法
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702489
Hisashi Igarashi, Yuichi Tanaka, Madoka Hasegawa, Shigeo Kato
In this paper, we propose a data embedding method into still images based on Vector Quantization (VQ). In recent years, several VQ-based data embedding methods have been proposed. For examle, ‘Mean Gray-Level Embedding method (MGLE)’ are ‘Pair wise Nearest-Neighbor Embedding method (PNNE)’ are simple, but not sufficiently effective. Meanwhile, an efficient adaptive data hiding method called ‘Adaptive Clustering Embedding method (ACE)’ was proposed, but is somewhat complicated because the VQ indices have to be adaptively clustered in the embedding process. In our proposed method, output vectors are considered as nodes, and nodes are linked as a tree structure and information is embedded by using some of linked vectors. The simulation results show that our proposed method indicates higher SNR than the conventional methods under the same amounts of embedded data.
本文提出了一种基于矢量量化(VQ)的静态图像数据嵌入方法。近年来,人们提出了几种基于vq的数据嵌入方法。例如,“平均灰度嵌入法(MGLE)”和“成对最近邻嵌入法(PNNE)”都很简单,但不够有效。同时,提出了一种高效的自适应数据隐藏方法“自适应聚类嵌入法”(ACE),但由于嵌入过程中需要对VQ指标进行自适应聚类,该方法比较复杂。在我们提出的方法中,输出向量被认为是节点,节点被链接成一个树结构,并通过一些链接向量嵌入信息。仿真结果表明,在相同的嵌入数据量下,该方法比传统方法具有更高的信噪比。
{"title":"VQ based data hiding method for still images by tree-structured links","authors":"Hisashi Igarashi, Yuichi Tanaka, Madoka Hasegawa, Shigeo Kato","doi":"10.1109/PCS.2010.5702489","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702489","url":null,"abstract":"In this paper, we propose a data embedding method into still images based on Vector Quantization (VQ). In recent years, several VQ-based data embedding methods have been proposed. For examle, ‘Mean Gray-Level Embedding method (MGLE)’ are ‘Pair wise Nearest-Neighbor Embedding method (PNNE)’ are simple, but not sufficiently effective. Meanwhile, an efficient adaptive data hiding method called ‘Adaptive Clustering Embedding method (ACE)’ was proposed, but is somewhat complicated because the VQ indices have to be adaptively clustered in the embedding process. In our proposed method, output vectors are considered as nodes, and nodes are linked as a tree structure and information is embedded by using some of linked vectors. The simulation results show that our proposed method indicates higher SNR than the conventional methods under the same amounts of embedded data.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132388960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Direction-adaptive hierarchical decomposition for image coding 方向自适应分层分解图像编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702566
Tomokazu Murakami, Keita Takahashi, T. Naemura
A new model of decomposing an image hierarchically into direction-adaptive subbands using pixel-wise direction estimation is presented. For each decomposing operation, an input image is divided into two parts: a base image subsampled from the input image and subband components. The subband components consist of residuals of estimating the pixels skipped through the subsampling, which ensures the invertibility of the decomposition. The estimation is performed in a direction-adaptive way, whose optimal direction is determined by a L1 norm criterion for each pixel, aiming to achieve good energy compaction that is suitable for image coding. Furthermore, since the L1 norms are obtained from the base image alone, we do not need to retain the directional information explicitly, which is another advantage of our model. Experimental results show that the proposed model can achieve lower entropy than conventional Haar or D5/3 discrete wavelet transform in case of lossless coding.
提出了一种利用逐像素方向估计将图像分层分解为方向自适应子带的新模型。对于每个分解操作,输入图像被分成两个部分:从输入图像中抽取的基础图像和子带分量。子带分量由估计跳过子采样的像素的残差组成,保证了分解的可逆性。采用方向自适应的方式进行估计,通过每个像素的L1范数准则确定其最优方向,以获得适合图像编码的良好能量压缩。此外,由于L1范数仅从基础图像中获得,我们不需要明确保留方向信息,这是我们模型的另一个优势。实验结果表明,在无损编码的情况下,该模型比传统Haar或D5/3离散小波变换获得更低的熵值。
{"title":"Direction-adaptive hierarchical decomposition for image coding","authors":"Tomokazu Murakami, Keita Takahashi, T. Naemura","doi":"10.1109/PCS.2010.5702566","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702566","url":null,"abstract":"A new model of decomposing an image hierarchically into direction-adaptive subbands using pixel-wise direction estimation is presented. For each decomposing operation, an input image is divided into two parts: a base image subsampled from the input image and subband components. The subband components consist of residuals of estimating the pixels skipped through the subsampling, which ensures the invertibility of the decomposition. The estimation is performed in a direction-adaptive way, whose optimal direction is determined by a L1 norm criterion for each pixel, aiming to achieve good energy compaction that is suitable for image coding. Furthermore, since the L1 norms are obtained from the base image alone, we do not need to retain the directional information explicitly, which is another advantage of our model. Experimental results show that the proposed model can achieve lower entropy than conventional Haar or D5/3 discrete wavelet transform in case of lossless coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132509721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Enhanced region-based adaptive interpolation filter 增强的基于区域的自适应插值滤波器
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702554
Shohei Matsuo, Y. Bandoh, Seishi Takamura, H. Jozawa
Motion compensation with quarter-pel accuracy was added to H.264/AVC to improve the coding efficiency of images exhibiting fractional-pel movement. To enlarge the reference pictures, a fixed 6-tap filter is used. However, the values of the filter coefficients are constant regardless of the characteristic of the input video. An improved interpolation filter, called the Adaptive Interpolation Filter (AIF), that optimizes the filter coefficients on a frame-by-frame basis was proposed to solve the problem. However, when the image is divided into multiple regions, each of which has different characteristics, the coding efficiency could be futher improved by performing optimization on a region-by-region basis. Therefore, we propose a Region-Based AIF (RBAIF) that takes account of image locality. Simulations show that RBAIF offers about 0.43 point higher coding gain than the conventional AIF.
在H.264/AVC中加入四分之一像素精度的运动补偿,提高了分数像素运动图像的编码效率。为了放大参考图片,使用了一个固定的6抽头滤波器。然而,无论输入视频的特性如何,滤波器系数的值都是恒定的。为了解决这一问题,提出了一种改进的插值滤波器,称为自适应插值滤波器(AIF),它在逐帧的基础上优化滤波器系数。然而,当图像被划分为多个区域时,每个区域都有不同的特征,可以通过逐区域优化来进一步提高编码效率。因此,我们提出了一种考虑图像局部性的基于区域的AIF (RBAIF)。仿真结果表明,RBAIF的编码增益比传统的AIF高0.43个点。
{"title":"Enhanced region-based adaptive interpolation filter","authors":"Shohei Matsuo, Y. Bandoh, Seishi Takamura, H. Jozawa","doi":"10.1109/PCS.2010.5702554","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702554","url":null,"abstract":"Motion compensation with quarter-pel accuracy was added to H.264/AVC to improve the coding efficiency of images exhibiting fractional-pel movement. To enlarge the reference pictures, a fixed 6-tap filter is used. However, the values of the filter coefficients are constant regardless of the characteristic of the input video. An improved interpolation filter, called the Adaptive Interpolation Filter (AIF), that optimizes the filter coefficients on a frame-by-frame basis was proposed to solve the problem. However, when the image is divided into multiple regions, each of which has different characteristics, the coding efficiency could be futher improved by performing optimization on a region-by-region basis. Therefore, we propose a Region-Based AIF (RBAIF) that takes account of image locality. Simulations show that RBAIF offers about 0.43 point higher coding gain than the conventional AIF.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132610476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
28th Picture Coding Symposium
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1