This paper describes a scalable extension of the High Efficiency Video Coding (HEVC) standard that supports different color gamuts in an enhancement and base layer. Here, the emphasis is on scenarios with BT.2020 color gamut in an enhancement layer and BT.709 color gamut in the base layer. This is motivated by a need to provide content for both high definition and ultra-high definition devices in the near future. The paper describes a method for predicting the enhancement layer samples from a decoded base layer using a series of multiplies and adds to account for both color gamut and bit-depth changes. Results show an improvement in coding efficiency between 65% and 84% for luma (57% and 85% for chroma) compared to simulcast in quality (SNR) scalable coding.
{"title":"Color Gamut Scalable Video Coding","authors":"L. Kerofsky, C. A. Segall, Seung-Hwan Kim","doi":"10.1109/DCC.2013.29","DOIUrl":"https://doi.org/10.1109/DCC.2013.29","url":null,"abstract":"This paper describes a scalable extension of the High Efficiency Video Coding (HEVC) standard that supports different color gamuts in an enhancement and base layer. Here, the emphasis is on scenarios with BT.2020 color gamut in an enhancement layer and BT.709 color gamut in the base layer. This is motivated by a need to provide content for both high definition and ultra-high definition devices in the near future. The paper describes a method for predicting the enhancement layer samples from a decoded base layer using a series of multiplies and adds to account for both color gamut and bit-depth changes. Results show an improvement in coding efficiency between 65% and 84% for luma (57% and 85% for chroma) compared to simulcast in quality (SNR) scalable coding.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128075669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper presents an innovative and efficient approach to code compression. Our method reduces code size by up to 32.6% and 31.9% (including all extra costs) respectively, for ARM and MIPS processor, and presents an improvement of almost 7% over the traditional Huffman method. We performed simulations and analyzes, using the applications from benchmark MiBench. In spite of these experiments, our method is orthogonal to approaches that take into account the particularities of a given instruction set architecture, becoming an independent method for any specific architecture.
{"title":"Multi-Level Dictionary Used in Code Compression for Embedded Systems","authors":"W. R. A. Dias, E. Moreno","doi":"10.1109/DCC.2013.69","DOIUrl":"https://doi.org/10.1109/DCC.2013.69","url":null,"abstract":"This paper presents an innovative and efficient approach to code compression. Our method reduces code size by up to 32.6% and 31.9% (including all extra costs) respectively, for ARM and MIPS processor, and presents an improvement of almost 7% over the traditional Huffman method. We performed simulations and analyzes, using the applications from benchmark MiBench. In spite of these experiments, our method is orthogonal to approaches that take into account the particularities of a given instruction set architecture, becoming an independent method for any specific architecture.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132092805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
John P. T. Moore, Antonio D. Kheirkhahzadeh, Jiva N. Bagale
Our compression technique is an abstraction of Packed Encoding Rules and has been implemented by the Packed objects structured data compression tool. Rather than trying to support a complex standard we instead describe a very simple technique which allows us to implement a very light-weight encoder capable of compressing structured data represented in XML. We call this work Integer Encoding Rules (IER). The technique is based on a simple mapping of data values belonging to a set of data types to a series of integer values. The data values come from XML data and the data types come from XML Schema.
{"title":"Domain-Specific XML Compression","authors":"John P. T. Moore, Antonio D. Kheirkhahzadeh, Jiva N. Bagale","doi":"10.1109/DCC.2013.90","DOIUrl":"https://doi.org/10.1109/DCC.2013.90","url":null,"abstract":"Our compression technique is an abstraction of Packed Encoding Rules and has been implemented by the Packed objects structured data compression tool. Rather than trying to support a complex standard we instead describe a very simple technique which allows us to implement a very light-weight encoder capable of compressing structured data represented in XML. We call this work Integer Encoding Rules (IER). The technique is based on a simple mapping of data values belonging to a set of data types to a series of integer values. The data values come from XML data and the data types come from XML Schema.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129799846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Martínez-Rach, O. López, P. Piñol, Manuel P. Malumbres
Depth Image Base Rendering (DIBR) is an effective approach for 3D-TV, however quality and time consistence of the depth map is a problem in this field. Our intermediate solution between Intra and Inter encoders is able to cope with the quality and time consistency of the captured depth map info. Our encoder achieves the same visual quality than H264/AVC and x264 in Intra mode reducing coding delays.
深度底图绘制(deep Image Base Rendering, DIBR)是一种有效的3d电视图像绘制方法,但深度图的质量和时间一致性一直是该领域存在的问题。我们在Intra和Inter编码器之间的中间解决方案能够处理捕获深度图信息的质量和时间一致性。我们的编码器在Intra模式下实现了与H264/AVC和x264相同的视觉质量,减少了编码延迟。
{"title":"3D Wavelet Encoder for Depth Map Data Compression","authors":"M. Martínez-Rach, O. López, P. Piñol, Manuel P. Malumbres","doi":"10.1109/DCC.2013.88","DOIUrl":"https://doi.org/10.1109/DCC.2013.88","url":null,"abstract":"Depth Image Base Rendering (DIBR) is an effective approach for 3D-TV, however quality and time consistence of the depth map is a problem in this field. Our intermediate solution between Intra and Inter encoders is able to cope with the quality and time consistency of the captured depth map info. Our encoder achieves the same visual quality than H264/AVC and x264 in Intra mode reducing coding delays.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"02 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129934218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Leandro Jimenez-Rodriguez, Francesc Aulí Llinàs, M. Marcellin, J. Serra-Sagristà
Visually lossless coding is a method through which an image is coded with numerical losses that are not noticeable by visual inspection. Contrary to numerically lossless coding, visually lossless coding can achieve high compression ratios. In general, visually lossless coding is approached from the point of view of the encoder, i.e., as a procedure devised to generate a compressed code stream from an original image. If an image has already been encoded to a very high fidelity (higher than visually lossless - perhaps even numerically lossless), it is not straightforward to create a just visually lossless version without fully re-encoding the image. However, for large repositories, re-encoding may not be a suitable option. A visually lossless decoder might be useful to decode, or to parse and transmit, only the data needed for visually lossless reconstruction. This work introduces a decoder for JPEG 2000 code streams that identifies and decodes the minimum amount of information needed to produce a visually lossless image. The main insights behind the proposed method are to estimate the variance of the code blocks before the decoding procedure, and to determine the visibility thresholds employing a well-known model from the literature. The main advantages are faster decoding and the possibility to transmit visually lossless images employing minimal bit rates.
{"title":"Visually Lossless JPEG 2000 Decoder","authors":"Leandro Jimenez-Rodriguez, Francesc Aulí Llinàs, M. Marcellin, J. Serra-Sagristà","doi":"10.1109/DCC.2013.25","DOIUrl":"https://doi.org/10.1109/DCC.2013.25","url":null,"abstract":"Visually lossless coding is a method through which an image is coded with numerical losses that are not noticeable by visual inspection. Contrary to numerically lossless coding, visually lossless coding can achieve high compression ratios. In general, visually lossless coding is approached from the point of view of the encoder, i.e., as a procedure devised to generate a compressed code stream from an original image. If an image has already been encoded to a very high fidelity (higher than visually lossless - perhaps even numerically lossless), it is not straightforward to create a just visually lossless version without fully re-encoding the image. However, for large repositories, re-encoding may not be a suitable option. A visually lossless decoder might be useful to decode, or to parse and transmit, only the data needed for visually lossless reconstruction. This work introduces a decoder for JPEG 2000 code streams that identifies and decodes the minimum amount of information needed to produce a visually lossless image. The main insights behind the proposed method are to estimate the variance of the code blocks before the decoding procedure, and to determine the visibility thresholds employing a well-known model from the literature. The main advantages are faster decoding and the possibility to transmit visually lossless images employing minimal bit rates.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130601988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Traditional rate-distortion theory is focused on how to best encode a signal using as few bits as possible and incurring as low a distortion as possible. However, very often, the goal of transmission is to extract specific information from the signal at the receiving end, and the distortion should be measured on that extracted information. In this paper we examine the problem of encoding signals such that sufficient information is preserved about their pair wise distances. For that goal, we consider randomized embeddings as an encoding mechanism and provide a framework to analyze their performance. We also propose the recently developed universal quantized embeddings as a solution to that problem and experimentally demonstrate that, in image retrieval experiments, universal embedding can achieve up to 25% rate reduction over the state of the art.
{"title":"Efficient Coding of Signal Distances Using Universal Quantized Embeddings","authors":"P. Boufounos, S. Rane","doi":"10.1109/DCC.2013.33","DOIUrl":"https://doi.org/10.1109/DCC.2013.33","url":null,"abstract":"Traditional rate-distortion theory is focused on how to best encode a signal using as few bits as possible and incurring as low a distortion as possible. However, very often, the goal of transmission is to extract specific information from the signal at the receiving end, and the distortion should be measured on that extracted information. In this paper we examine the problem of encoding signals such that sufficient information is preserved about their pair wise distances. For that goal, we consider randomized embeddings as an encoding mechanism and provide a framework to analyze their performance. We also propose the recently developed universal quantized embeddings as a solution to that problem and experimentally demonstrate that, in image retrieval experiments, universal embedding can achieve up to 25% rate reduction over the state of the art.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121363607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Guang Y. Zhang, Abdelrahman Abdelazim, S. Mein, M. Varley, D. Ait-Boudaoud
Summary form only given. Multiple video cameras are used to capture the same scene simultaneously to acquire the multiview view coding data, obviously, over-large data will affect the coding efficiency. Due to the video data is acquired from the same scene, the inter-view similarities between adjacent camera views are exploited for efficient compression. Generally, the same objects with different viewpoints are shown on adjacent views. On the other hand, containing objects at different depth planes, and therefore perfect correlation over the entire image area will never occur. Additionally, the scene complexity and the differences in brightness and color between the video of the individual cameras will also affect the current block to find its best match in the inter-view reference picture. Consequently, the temporal-view reference picture is referred more frequently. In order to gain the compression efficiency, it is a core part to disable the unnecessary inter-view reference. The idea of this paper is to exploit the phase correlation to estimate the dependencies between the inter-view reference and the current picture. If the two frames with low correlation, the inter-view reference frame will be disabled. In addition, this approach works only on non-anchor pictures. Experimental results show that the proposed algorithm can save 16% computational complexity on average, with negligible loss of quality and bit rate. The phase correlation process only takes up 0.1% of the whole process.
{"title":"Inter-view Reference Frame Selection in Multi-view Video Coding","authors":"Guang Y. Zhang, Abdelrahman Abdelazim, S. Mein, M. Varley, D. Ait-Boudaoud","doi":"10.1109/DCC.2013.113","DOIUrl":"https://doi.org/10.1109/DCC.2013.113","url":null,"abstract":"Summary form only given. Multiple video cameras are used to capture the same scene simultaneously to acquire the multiview view coding data, obviously, over-large data will affect the coding efficiency. Due to the video data is acquired from the same scene, the inter-view similarities between adjacent camera views are exploited for efficient compression. Generally, the same objects with different viewpoints are shown on adjacent views. On the other hand, containing objects at different depth planes, and therefore perfect correlation over the entire image area will never occur. Additionally, the scene complexity and the differences in brightness and color between the video of the individual cameras will also affect the current block to find its best match in the inter-view reference picture. Consequently, the temporal-view reference picture is referred more frequently. In order to gain the compression efficiency, it is a core part to disable the unnecessary inter-view reference. The idea of this paper is to exploit the phase correlation to estimate the dependencies between the inter-view reference and the current picture. If the two frames with low correlation, the inter-view reference frame will be disabled. In addition, this approach works only on non-anchor pictures. Experimental results show that the proposed algorithm can save 16% computational complexity on average, with negligible loss of quality and bit rate. The phase correlation process only takes up 0.1% of the whole process.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121723810","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Toshiya Tanaka, T. I., Shunsuke Inenaga, H. Bannai, M. Takeda
The convolution between a text string S of length N and a pattern string P of length m can be computed in O(N log m) time by FFT. It is known that various types of approximate string matching problems are reducible to convolution. In this paper, we assume that the input text string is given in a compressed form, as a straight-line program (SLP), which is a context free grammar in the Chomsky normal form that derives a single string. Given an SLP S of size n describing a text S of length N, and an uncompressed pattern P of length m, we present a simple O(nm log m)-time algorithm to compute the convolution between S and P. We then show that this can be improved to O(min{nm, N - α} log m) time, where α ≥ 0 is a value that represents the amount of redundancy that the SLP captures with respect to the length-m substrings. The key of the improvement is our new algorithm that computes the convolution between a trie of size r and a pattern string P of length m in O(r log m) time.
{"title":"Computing Convolution on Grammar-Compressed Text","authors":"Toshiya Tanaka, T. I., Shunsuke Inenaga, H. Bannai, M. Takeda","doi":"10.1109/DCC.2013.53","DOIUrl":"https://doi.org/10.1109/DCC.2013.53","url":null,"abstract":"The convolution between a text string S of length N and a pattern string P of length m can be computed in O(N log m) time by FFT. It is known that various types of approximate string matching problems are reducible to convolution. In this paper, we assume that the input text string is given in a compressed form, as a straight-line program (SLP), which is a context free grammar in the Chomsky normal form that derives a single string. Given an SLP S of size n describing a text S of length N, and an uncompressed pattern P of length m, we present a simple O(nm log m)-time algorithm to compute the convolution between S and P. We then show that this can be improved to O(min{nm, N - α} log m) time, where α ≥ 0 is a value that represents the amount of redundancy that the SLP captures with respect to the length-m substrings. The key of the improvement is our new algorithm that computes the convolution between a trie of size r and a pattern string P of length m in O(r log m) time.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"74 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124752674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Summary form only given. A Markov decision process (MDP) is defined by a state space, action space, transition model, and reward model. The objective is to maximize accumulation of reward over time. Solutions can be found through dynamic programming, which generally involves discretization, resulting in significant memory and computational requirements. Although computer clusters can be used to solve large problems, many applications require that solutions be executed on less capable hardware. We explored a general method for compressing solutions in a way that preserves fast random-access lookups. The method was applied to an MDP for an aircraft collision avoidance system. In our problem, S consists of aircraft positions and velocities and A consists of resolution advisories provided by the collision avoidance system, with S > 1.5 x 106, and A = 10. The solution to an MDP can be represented by an |S| x |A| matrix specifying Q*(s,a), the expected return of the optimal strategy from s after executing action a. Since, on average, only 6.6 actions are available from every state in our problem, it is more efficient to use a sparse representation consisting of an array of the permissible values of Q*, organized into into variable-length blocks with one block per state. An index provides offsets into this Q* array corresponding to the block boundaries, and an action array lists the actions available from each state. The values for Q* are stored using a 32-bit floating point representation, resulting in 534 MB for the three arrays associated with the sparse representation. Our method first converts to a 16-bit half-precision representation, sorts the state-action values within each block, adjusts the action array appropriately, and then removes redundant blocks. Although LZMA has a better compression ratio, it does not support real-time random access decompression. The behavior of the proposed method was demonstrated in simulation with negligible impact on safety and operational performance metrics. Although this compression methodology was demonstrated on related MDPs with similar compression ratios, further work will apply this technique to other domains.
只提供摘要形式。马尔可夫决策过程由状态空间、动作空间、转移模型和奖励模型定义。目标是随着时间的推移最大化奖励的积累。可以通过动态规划找到解决方案,这通常涉及离散化,导致大量内存和计算需求。尽管计算机集群可用于解决大型问题,但许多应用程序要求在性能较差的硬件上执行解决方案。我们探索了一种压缩解决方案的通用方法,以保持快速随机访问查找。将该方法应用于某型飞机避碰系统的MDP中。在我们的问题中,S为飞机位置和速度,A为避碰系统提供的解决建议,S > 1.5 x 106, A = 10。MDP的解决方案可以用一个指定Q*(S, A)的|S| x |A|矩阵来表示,Q*(S, A)是执行动作A后从S得到的最优策略的预期回报。因为,在我们的问题中,每个状态平均只有6.6个动作可用,所以使用由Q*的允许值数组组成的稀疏表示更有效,这些值被组织成可变长度的块,每个状态一个块。索引提供了与块边界相对应的Q*数组的偏移量,动作数组列出了每个状态中可用的动作。Q*的值使用32位浮点表示进行存储,因此与稀疏表示相关联的三个数组占用534 MB。我们的方法首先转换为16位半精度表示,对每个块中的状态-动作值进行排序,适当地调整动作数组,然后删除多余的块。虽然LZMA具有更好的压缩比,但它不支持实时随机访问解压缩。所提出的方法的行为在仿真中得到了证明,对安全和运行性能指标的影响可以忽略不计。尽管这种压缩方法在具有相似压缩比的相关mdp上得到了证明,但进一步的工作将把这种技术应用于其他领域。
{"title":"Compression of Optimal Value Functions for Markov Decision Processes","authors":"Mykel J. Kochenderfer, Nicholas Monath","doi":"10.1109/DCC.2013.81","DOIUrl":"https://doi.org/10.1109/DCC.2013.81","url":null,"abstract":"Summary form only given. A Markov decision process (MDP) is defined by a state space, action space, transition model, and reward model. The objective is to maximize accumulation of reward over time. Solutions can be found through dynamic programming, which generally involves discretization, resulting in significant memory and computational requirements. Although computer clusters can be used to solve large problems, many applications require that solutions be executed on less capable hardware. We explored a general method for compressing solutions in a way that preserves fast random-access lookups. The method was applied to an MDP for an aircraft collision avoidance system. In our problem, S consists of aircraft positions and velocities and A consists of resolution advisories provided by the collision avoidance system, with S > 1.5 x 106, and A = 10. The solution to an MDP can be represented by an |S| x |A| matrix specifying Q*(s,a), the expected return of the optimal strategy from s after executing action a. Since, on average, only 6.6 actions are available from every state in our problem, it is more efficient to use a sparse representation consisting of an array of the permissible values of Q*, organized into into variable-length blocks with one block per state. An index provides offsets into this Q* array corresponding to the block boundaries, and an action array lists the actions available from each state. The values for Q* are stored using a 32-bit floating point representation, resulting in 534 MB for the three arrays associated with the sparse representation. Our method first converts to a 16-bit half-precision representation, sorts the state-action values within each block, adjusts the action array appropriately, and then removes redundant blocks. Although LZMA has a better compression ratio, it does not support real-time random access decompression. The behavior of the proposed method was demonstrated in simulation with negligible impact on safety and operational performance metrics. Although this compression methodology was demonstrated on related MDPs with similar compression ratios, further work will apply this technique to other domains.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129693578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}