Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702550
Shuliang Lin, Yuichiro Sawa, Norishige Fukushima, Y. Ishibashi
This paper analyzes the influences of frame delay and packet loss on stereoscopic vision when stereoscopic video is transferred over a IP network. We employ live action videos which are transferred to a head-mount-display (HMD) and do the assessment on stereoscopic perception. As a result, we found that speed and movement direction of the attention object play a great role on the deterioration when frame delay and packet loss exist.
{"title":"Influences of frame delay and packet loss between left and right frames in stereoscopic video communications","authors":"Shuliang Lin, Yuichiro Sawa, Norishige Fukushima, Y. Ishibashi","doi":"10.1109/PCS.2010.5702550","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702550","url":null,"abstract":"This paper analyzes the influences of frame delay and packet loss on stereoscopic vision when stereoscopic video is transferred over a IP network. We employ live action videos which are transferred to a head-mount-display (HMD) and do the assessment on stereoscopic perception. As a result, we found that speed and movement direction of the attention object play a great role on the deterioration when frame delay and packet loss exist.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115777503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702541
G. Chan, Jheng-Ping Lin, A. Tang
The high computational complexity of multi-view video codecs makes it necessary to speed up for their realization in consumer electronics. Since fast encoding algorithms are expected to adapt to different video sequences, this paper proposes a fast algorithm that consists of fast mode decision and fast disparity estimation for multi-view video coding. The fast mode decision algorithm applies to both temporal and inter-view predictions. The candidates for mode decision are reduced based on a set of thresholds. Differ from the previous fast mode decision algorithms for MVC, this scheme determines the thresholds according to the online statistical analysis of motion and disparity costs of the first GOP in each view. Since the inter-view prediction is time consuming, we propose a fast disparity estimation algorithm to save encoding time. Experimental results show that our proposed scheme reduces the computational complexity significantly with negligible degradation of coding efficiency.
{"title":"On-line statistical analysis based fast mode decision for multi-view video coding","authors":"G. Chan, Jheng-Ping Lin, A. Tang","doi":"10.1109/PCS.2010.5702541","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702541","url":null,"abstract":"The high computational complexity of multi-view video codecs makes it necessary to speed up for their realization in consumer electronics. Since fast encoding algorithms are expected to adapt to different video sequences, this paper proposes a fast algorithm that consists of fast mode decision and fast disparity estimation for multi-view video coding. The fast mode decision algorithm applies to both temporal and inter-view predictions. The candidates for mode decision are reduced based on a set of thresholds. Differ from the previous fast mode decision algorithms for MVC, this scheme determines the thresholds according to the online statistical analysis of motion and disparity costs of the first GOP in each view. Since the inter-view prediction is time consuming, we propose a fast disparity estimation algorithm to save encoding time. Experimental results show that our proposed scheme reduces the computational complexity significantly with negligible degradation of coding efficiency.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128866064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702436
T. Komatsu, Yasutaka Ueda, T. Saito
Alter, Durand and Froment introduced the total-variation (TV) minimization approach to the artifact-free JPEG decoding, which is referred to as the ADF decoding method [1]. They formulated the decoding problem as the constrained TV restoration problem, in which the TV seminorm of its restored color image is minimized under the constraint that each DCT coefficient of the restored color image should be in the quantization interval of its corresponding DCT coefficient of the JPEG-compressed data. This paper proposes a new restoration approach to the JPEG decoding. Instead of the TV regularization, our new JPEG-decoding method employs a shrinkage operation in the redundant DCT domain, to mitigate degradations caused by the JPEG coding. Our new method not only can selectively suppress ringing artifacts near color edges, but also can efficiently eliminate blocking artifacts in originally smoothly-varying image regions, where the blocking artifacts are very noticeable. Through decoding simulations, we experimentally show that our new decoding method can reduce JPEG-coding artifacts more effectively than the ADF decoding method.
Alter, Durand and Froment引入了全变差(total-variation, TV)最小化方法来实现无伪影的JPEG解码,称为ADF解码方法[1]。他们将解码问题表述为约束电视还原问题,即在还原的彩色图像的每个DCT系数应在jpeg压缩数据的相应DCT系数的量化区间内的约束下,将其还原的彩色图像的电视半形态最小化。本文提出了一种新的JPEG解码复原方法。我们的新JPEG解码方法采用冗余DCT域的收缩操作来代替TV正则化,以减轻JPEG编码引起的降级。该方法不仅可以选择性地抑制颜色边缘附近的环状伪影,而且可以有效地消除原本平滑变化的图像区域的阻塞伪影,在这些区域阻塞伪影非常明显。通过解码仿真,我们实验证明了我们的解码方法能比ADF解码方法更有效地减少jpeg编码伪影。
{"title":"Super-resolution decoding of JPEG-compressed image data with the shrinkage in the redundant DCT domain","authors":"T. Komatsu, Yasutaka Ueda, T. Saito","doi":"10.1109/PCS.2010.5702436","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702436","url":null,"abstract":"Alter, Durand and Froment introduced the total-variation (TV) minimization approach to the artifact-free JPEG decoding, which is referred to as the ADF decoding method [1]. They formulated the decoding problem as the constrained TV restoration problem, in which the TV seminorm of its restored color image is minimized under the constraint that each DCT coefficient of the restored color image should be in the quantization interval of its corresponding DCT coefficient of the JPEG-compressed data. This paper proposes a new restoration approach to the JPEG decoding. Instead of the TV regularization, our new JPEG-decoding method employs a shrinkage operation in the redundant DCT domain, to mitigate degradations caused by the JPEG coding. Our new method not only can selectively suppress ringing artifacts near color edges, but also can efficiently eliminate blocking artifacts in originally smoothly-varying image regions, where the blocking artifacts are very noticeable. Through decoding simulations, we experimentally show that our new decoding method can reduce JPEG-coding artifacts more effectively than the ADF decoding method.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"224 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131262123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702586
Jie Xu, Jianwei Ma, Dongming Zhang, Yongdong Zhang, Shouxun Lin
We propose a compressive video sensing scheme based on user attention model (UAM) for real video sequences acquisition. In this work, for every group of consecutive video frames, we set the first frame as reference frame and build a UAM with visual rhythm analysis (VRA) to automatically determine region-of-interest (ROI) for non-reference frames. The determined ROI usually has significant movement and attracts more attention. Each frame of the video sequence is divided into non-overlapping blocks of 16×16 pixel size. Compressive video sampling is conducted in a block-by-block manner on each frame through a single operator and in a whole region manner on the ROIs through a different operator. Our video reconstruction algorithm involves alternating direction l1 — norm minimization algorithm (ADM) for the frame difference of non-ROI blocks and minimum total-variance (TV) method for the ROIs. Experimental results showed that our method could significantly enhance the quality of reconstructed video and reduce the errors accumulated during the reconstruction.
{"title":"Compressive video sensing based on user attention model","authors":"Jie Xu, Jianwei Ma, Dongming Zhang, Yongdong Zhang, Shouxun Lin","doi":"10.1109/PCS.2010.5702586","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702586","url":null,"abstract":"We propose a compressive video sensing scheme based on user attention model (UAM) for real video sequences acquisition. In this work, for every group of consecutive video frames, we set the first frame as reference frame and build a UAM with visual rhythm analysis (VRA) to automatically determine region-of-interest (ROI) for non-reference frames. The determined ROI usually has significant movement and attracts more attention. Each frame of the video sequence is divided into non-overlapping blocks of 16×16 pixel size. Compressive video sampling is conducted in a block-by-block manner on each frame through a single operator and in a whole region manner on the ROIs through a different operator. Our video reconstruction algorithm involves alternating direction l1 — norm minimization algorithm (ADM) for the frame difference of non-ROI blocks and minimum total-variance (TV) method for the ROIs. Experimental results showed that our method could significantly enhance the quality of reconstructed video and reduce the errors accumulated during the reconstruction.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115708671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702447
A. Yoshitaka, Terumasa Hyoudou
In the area of retrieving image databases, one of the promising approaches is to retrieve it by specifying image example. However, specifying a single image example is not always sufficient to get satisfactory result, since one image example does not give comprehensive ranges of values that reflect the various aspects of the object to be retrieved. In this paper, we propose a method of retrieving images by specifying multiple image examples that is designed for retrieving sign boards. Features of color, shape, and spatial relation of color regions are extracted from example images, and they are clustered so as to obtain proper range of values. Compared with QBE systems that accept only a single image as the query condition, MIERS (Multi-Image Example-based Retrieval System) returns better retrieval result, where the experimental result showed that specifying more examples helps to improve recall with little deterioration of precision.
{"title":"Content-based retrieval by multiple image examples for sign board retrieval","authors":"A. Yoshitaka, Terumasa Hyoudou","doi":"10.1109/PCS.2010.5702447","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702447","url":null,"abstract":"In the area of retrieving image databases, one of the promising approaches is to retrieve it by specifying image example. However, specifying a single image example is not always sufficient to get satisfactory result, since one image example does not give comprehensive ranges of values that reflect the various aspects of the object to be retrieved. In this paper, we propose a method of retrieving images by specifying multiple image examples that is designed for retrieving sign boards. Features of color, shape, and spatial relation of color regions are extracted from example images, and they are clustered so as to obtain proper range of values. Compared with QBE systems that accept only a single image as the query condition, MIERS (Multi-Image Example-based Retrieval System) returns better retrieval result, where the experimental result showed that specifying more examples helps to improve recall with little deterioration of precision.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114920988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702553
Bumshik Lee, Jae-il Kim, Sangsoo Ahn, Munchurl Kim, Hui-Yong Kim, Jong-Ho Kim, J. Choi
In this paper, a rate-distortion optimized variable block transform coding scheme is proposed based on a hierarchical structured transform for macroblock (MB) coding with a set of the order-4 and −8 integer cosine transform (ICT) kernels of H.264/AVC as well as a new order-16 ICT kernel. The set of order-4, −8 and −16 ICT kernels are applied for inter-predictive coding in square (4×4, 8×8 or 16×16) or non-square (16×8 or 8×16) transform for each MB in a hierarchical structured manner. The proposed hierarchical variable-sized block transform scheme using the order-16 ICT kernel achieves significant bitrate reduction up to 15%, compared to the High profile of H.264/AVC. Even if the number of candidates for the transform types increases, the encoding time can be reduced to average 4–6% over the H.264/AVC
{"title":"A hierarchical variable-sized block transform coding scheme for coding efficiency improvement on H.264/AVC","authors":"Bumshik Lee, Jae-il Kim, Sangsoo Ahn, Munchurl Kim, Hui-Yong Kim, Jong-Ho Kim, J. Choi","doi":"10.1109/PCS.2010.5702553","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702553","url":null,"abstract":"In this paper, a rate-distortion optimized variable block transform coding scheme is proposed based on a hierarchical structured transform for macroblock (MB) coding with a set of the order-4 and −8 integer cosine transform (ICT) kernels of H.264/AVC as well as a new order-16 ICT kernel. The set of order-4, −8 and −16 ICT kernels are applied for inter-predictive coding in square (4×4, 8×8 or 16×16) or non-square (16×8 or 8×16) transform for each MB in a hierarchical structured manner. The proposed hierarchical variable-sized block transform scheme using the order-16 ICT kernel achieves significant bitrate reduction up to 15%, compared to the High profile of H.264/AVC. Even if the number of candidates for the transform types increases, the encoding time can be reduced to average 4–6% over the H.264/AVC","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114994715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702490
Danilo B. Graziosi, Nuno M. M. Rodrigues, C. Pagliari, E. Silva, S. Faria, Marcelo M. Perez, M. Carvalho
In this article we propose to compress depth maps using a coding scheme based on multiscale recurrent pattern matching and evaluate its impact on depth image based rendering (DIBR). Depth maps are usually converted into gray scale images and compressed like a conventional luminance signal. However, using traditional transform-based encoders to compress depth maps may result in undesired artifacts at sharp edges due to the quantization of high frequency coefficients. The Multidimensional Multiscale Parser (MMP) is a pattern matching-based encoder, that is able to preserve and efficiently encode high frequency patterns, such as edge information. This ability is critical for encoding depth map images. Experimental results for encoding depth maps show that MMP is much more efficient in a rate-distortion sense than standard image compression techniques such as JPEG2000 or H.264/AVC. In addition, the depth maps compressed with MMP generate reconstructed views with a higher quality than all other tested compression algorithms.
{"title":"Multiscale recurrent pattern matching approach for depth map coding","authors":"Danilo B. Graziosi, Nuno M. M. Rodrigues, C. Pagliari, E. Silva, S. Faria, Marcelo M. Perez, M. Carvalho","doi":"10.1109/PCS.2010.5702490","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702490","url":null,"abstract":"In this article we propose to compress depth maps using a coding scheme based on multiscale recurrent pattern matching and evaluate its impact on depth image based rendering (DIBR). Depth maps are usually converted into gray scale images and compressed like a conventional luminance signal. However, using traditional transform-based encoders to compress depth maps may result in undesired artifacts at sharp edges due to the quantization of high frequency coefficients. The Multidimensional Multiscale Parser (MMP) is a pattern matching-based encoder, that is able to preserve and efficiently encode high frequency patterns, such as edge information. This ability is critical for encoding depth map images. Experimental results for encoding depth maps show that MMP is much more efficient in a rate-distortion sense than standard image compression techniques such as JPEG2000 or H.264/AVC. In addition, the depth maps compressed with MMP generate reconstructed views with a higher quality than all other tested compression algorithms.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116451860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702502
Guolei Yang, Luhong Liang, Wen Gao
Fast stereoscopic video encoding becomes a highly desired technique because the stereoscopic video has been realizable for applications like TV broadcasting and consumer electronics. The stereoscopic video has high inter-view dependency subject to epipolar restriction, which can be used to reduce the encoding complexity. In this paper, we propose a fast inter-prediction mode selection algorithm for stereoscopic video encoding. Different from methods using disparity estimation, candidate modes are generated by sliding a window along the macro-block line restricted by the epipolar. Then the motion information is utilized to rectify the candidate modes. A selection failure handling algorithm is also proposed to preserve coding quality. The proposed algorithm is evaluated using independent H.264/AVC encoders for left and right views and can be extended to MVC. Experimental results show that encoding times of one view are reduced by 41.4% and 24.4% for HD and VGA videos respectively with little quality loss.
{"title":"An epipolar resticted inter-mode selection for stereoscopic video encoding","authors":"Guolei Yang, Luhong Liang, Wen Gao","doi":"10.1109/PCS.2010.5702502","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702502","url":null,"abstract":"Fast stereoscopic video encoding becomes a highly desired technique because the stereoscopic video has been realizable for applications like TV broadcasting and consumer electronics. The stereoscopic video has high inter-view dependency subject to epipolar restriction, which can be used to reduce the encoding complexity. In this paper, we propose a fast inter-prediction mode selection algorithm for stereoscopic video encoding. Different from methods using disparity estimation, candidate modes are generated by sliding a window along the macro-block line restricted by the epipolar. Then the motion information is utilized to rectify the candidate modes. A selection failure handling algorithm is also proposed to preserve coding quality. The proposed algorithm is evaluated using independent H.264/AVC encoders for left and right views and can be extended to MVC. Experimental results show that encoding times of one view are reduced by 41.4% and 24.4% for HD and VGA videos respectively with little quality loss.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"64 245 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125959075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702526
Z. Omar, N. Mitianoudis, T. Stathaki
This report documents in detail the research carried out by the author throughout his first year. The paper presents a novel method for fusing images in a domain concerning multiple sensors and modalities. Using Chebyshev polynomials as basis functions, the image is decomposed to perform fusion at feature level. Results show favourable performance compared to previous efforts on image fusion, namely ICA and DT-CWT, in noise affected images. The work presented here aims at providing a novel framework for future studies in image analysis and may introduce innovations in the fields of surveillance, medical imaging and remote sensing.
{"title":"Two-dimensional Chebyshev polynomials for image fusion","authors":"Z. Omar, N. Mitianoudis, T. Stathaki","doi":"10.1109/PCS.2010.5702526","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702526","url":null,"abstract":"This report documents in detail the research carried out by the author throughout his first year. The paper presents a novel method for fusing images in a domain concerning multiple sensors and modalities. Using Chebyshev polynomials as basis functions, the image is decomposed to perform fusion at feature level. Results show favourable performance compared to previous efforts on image fusion, namely ICA and DT-CWT, in noise affected images. The work presented here aims at providing a novel framework for future studies in image analysis and may introduce innovations in the fields of surveillance, medical imaging and remote sensing.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124745444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702435
M. Iwahashi, H. Kiya
In this report, permutation of order and sign of signals are introduced to avoid singular point problem of a reversible transform. When a transform is implemented in the lifting structure, it can be "reversible" in spite of rounding operations inside the transform. Therefore it has been applied to lossless coding of digital signals. However some coefficient values of the transform have singular points (SP). Around the SP, rounding errors are magnified to huge amount and the coding efficiency is decreased. In this report, we analyze the SP of a three point KLT for RGB color components of an image signal, and introduce permutation of order and sign of signals to avoid the SP problem. It was experimentally confirmed that the proposed method improved PSNR by approximately 15 [dB] comparing to the worst case.
{"title":"Avoidance of singular point in reversible KLT","authors":"M. Iwahashi, H. Kiya","doi":"10.1109/PCS.2010.5702435","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702435","url":null,"abstract":"In this report, permutation of order and sign of signals are introduced to avoid singular point problem of a reversible transform. When a transform is implemented in the lifting structure, it can be \"reversible\" in spite of rounding operations inside the transform. Therefore it has been applied to lossless coding of digital signals. However some coefficient values of the transform have singular points (SP). Around the SP, rounding errors are magnified to huge amount and the coding efficiency is decreased. In this report, we analyze the SP of a three point KLT for RGB color components of an image signal, and introduce permutation of order and sign of signals to avoid the SP problem. It was experimentally confirmed that the proposed method improved PSNR by approximately 15 [dB] comparing to the worst case.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128537544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}