Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702469
S. Yano
First, proper characteristics of stereoscopic or multi-view displays are described in the context of human visual perception. Next, the development of three-dimensional displays, matched to the function of human visual field, is described accounting for these points of stereoscopic and multi-view displays. These three-dimensional displays were developed based on the principle of light field reproduction, and the performances of each display are briefly described.
{"title":"Progress from stereoscopic to three-dimensional displays based on visual perception","authors":"S. Yano","doi":"10.1109/PCS.2010.5702469","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702469","url":null,"abstract":"First, proper characteristics of stereoscopic or multi-view displays are described in the context of human visual perception. Next, the development of three-dimensional displays, matched to the function of human visual field, is described accounting for these points of stereoscopic and multi-view displays. These three-dimensional displays were developed based on the principle of light field reproduction, and the performances of each display are briefly described.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124357804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702483
Yanwei Liu, S. Ci, Hui Tang
In 3D video communication, how to select the appropriate quantization parameter (QP) for depth map coding is very important for obtaining the optimal view synthesis quality. This paper first analyzes the depth uncertainty induced two kinds of view synthesis errors, namely the original depth error induced view synthesis error and the depth compression induced view synthesis error, and then proposes a quadratic model to characterize the relationship between the view synthesis quality and the depth quantization step size. The proposed model can find the inflexion point in the curve of the view synthesis quality with the increasing depth quantization step size. Experimental results show that, given the rate constraint for depth map, the proposed model can accurately find the optimal QP for depth map coding.
{"title":"View synthesis error analysis for selecting the optimal QP of depth map coding in 3D video application","authors":"Yanwei Liu, S. Ci, Hui Tang","doi":"10.1109/PCS.2010.5702483","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702483","url":null,"abstract":"In 3D video communication, how to select the appropriate quantization parameter (QP) for depth map coding is very important for obtaining the optimal view synthesis quality. This paper first analyzes the depth uncertainty induced two kinds of view synthesis errors, namely the original depth error induced view synthesis error and the depth compression induced view synthesis error, and then proposes a quadratic model to characterize the relationship between the view synthesis quality and the depth quantization step size. The proposed model can find the inflexion point in the curve of the view synthesis quality with the increasing depth quantization step size. Experimental results show that, given the rate constraint for depth map, the proposed model can accurately find the optimal QP for depth map coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115417021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702528
T. Saito, K. Ishikawa, Yasutaka Ueda, T. Komatsu
To remove signal-dependent noise of a digital color camera, we propose a new denoising method with our hard color-shrinkage in the tight-frame grouplet transform domain. The classic hard-shrinkage works well for monochrome-image denoising. To utilize inter-channel color dependence, a noisy image undergoes the color transformation from the RGB to the luminance-and-chrominance color space, and the luminance and the chrominance components are separately denoised; but this approach cannot cope with actual signal-dependent noise. To utilize the noise's signal-dependencies, we have constructed the hard color-shrinkage where the interchannel color dependence is directly utilized in the RGB color space. The hard color-shrinkage alleviates denoising artifacts, and improves picture quality of denoised images.
{"title":"Image denoising with hard color-shrinkage and grouplet transform","authors":"T. Saito, K. Ishikawa, Yasutaka Ueda, T. Komatsu","doi":"10.1109/PCS.2010.5702528","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702528","url":null,"abstract":"To remove signal-dependent noise of a digital color camera, we propose a new denoising method with our hard color-shrinkage in the tight-frame grouplet transform domain. The classic hard-shrinkage works well for monochrome-image denoising. To utilize inter-channel color dependence, a noisy image undergoes the color transformation from the RGB to the luminance-and-chrominance color space, and the luminance and the chrominance components are separately denoised; but this approach cannot cope with actual signal-dependent noise. To utilize the noise's signal-dependencies, we have constructed the hard color-shrinkage where the interchannel color dependence is directly utilized in the RGB color space. The hard color-shrinkage alleviates denoising artifacts, and improves picture quality of denoised images.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124906073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702514
Axel Lakus-Becker, Ka-Ming Leung, Zhonghua Ma
Side information plays a key role in the performance of a Distributed Video Coding (DVC) system. However, the generation of side information often relies on complex motion estimation/interpolation operation. The correlation between the source data and the side information, sometimes referred as virtual channel, is also very difficult to model accurately. In this paper, we propose a bitwise prediction error correction method to improve the quality of the side information during Wyner-Ziv decoding. Whenever a bit error is detected in a bit plane, the less significant bits of the corresponding pixel are adjusted to match the initial prediction. The proposed method has been evaluated using a pixel-domain DVC system and delivers a better coding performance with improved decoding quality and reduced bitrate.
{"title":"Bitwise prediction error correction for Distributed Video Coding","authors":"Axel Lakus-Becker, Ka-Ming Leung, Zhonghua Ma","doi":"10.1109/PCS.2010.5702514","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702514","url":null,"abstract":"Side information plays a key role in the performance of a Distributed Video Coding (DVC) system. However, the generation of side information often relies on complex motion estimation/interpolation operation. The correlation between the source data and the side information, sometimes referred as virtual channel, is also very difficult to model accurately. In this paper, we propose a bitwise prediction error correction method to improve the quality of the side information during Wyner-Ziv decoding. Whenever a bit error is detected in a bit plane, the less significant bits of the corresponding pixel are adjusted to match the initial prediction. The proposed method has been evaluated using a pixel-domain DVC system and delivers a better coding performance with improved decoding quality and reduced bitrate.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129749402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702488
Muhammad Majid, G. Abhayaratne
Scalable multiple description coding (SMDC) provides reliability and facility to truncate the descriptions according to the user rate-distortion requirements. In this paper we generalize the conditions of successive refinement of the side quantizer of a multiple description scalar quantizer that has overlapped quantizer cells generated by a modified linear index assignment matrix. We propose that the split or refinement factor for each of the refinement side quantizers should be greater than the maximum side quantizer bin spread and should not be integer multiples of each other for satisfying the SMDC distortion conditions and verify through simulation results on scalable multiple description image coding.
{"title":"Successive refinement of overlapped cell side quantizers for scalable multiple description coding","authors":"Muhammad Majid, G. Abhayaratne","doi":"10.1109/PCS.2010.5702488","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702488","url":null,"abstract":"Scalable multiple description coding (SMDC) provides reliability and facility to truncate the descriptions according to the user rate-distortion requirements. In this paper we generalize the conditions of successive refinement of the side quantizer of a multiple description scalar quantizer that has overlapped quantizer cells generated by a modified linear index assignment matrix. We propose that the split or refinement factor for each of the refinement side quantizers should be greater than the maximum side quantizer bin spread and should not be integer multiples of each other for satisfying the SMDC distortion conditions and verify through simulation results on scalable multiple description image coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128679768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702523
Lingchen Zhu, H. Xiong
Regarding orientation resolution as an isolated variable from scale, this paper proposes a dual (scale and orientation) multiresolution transform into scalable video coding (SVC) framework. By projecting 2D signals (textures and edges) onto nonuniformly divided orientation subspaces, the dual multiresolution SVC (DMSVC) can capture 2-D curve smoothness with less coefficients, and provide more flexible spatial decomposition structures than traditional wavelet-based SVC (WSVC). In the spatial decomposition module of DMSVC, the nonuniform directional distribution along scale of each frame is detected by phase congruency in the overcomplete wavelet domain. The corresponding orientational multi-resolution is achieved by nonuniform directional filter banks (NUDFB) which is fulfilled via a non-symmetric binary tree (NSBT) structured frequency division. The wavelet basis function in each scale is converted to an adaptive set of nonuniform directional basis by employing nonuniform directional filter banks. Experimental results validate a superior coding performance and visual quality over WSVC especially on those sequences full of directional edges and textures.
{"title":"Scalable video compression framework with adaptive multiresolution directional filter bank design","authors":"Lingchen Zhu, H. Xiong","doi":"10.1109/PCS.2010.5702523","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702523","url":null,"abstract":"Regarding orientation resolution as an isolated variable from scale, this paper proposes a dual (scale and orientation) multiresolution transform into scalable video coding (SVC) framework. By projecting 2D signals (textures and edges) onto nonuniformly divided orientation subspaces, the dual multiresolution SVC (DMSVC) can capture 2-D curve smoothness with less coefficients, and provide more flexible spatial decomposition structures than traditional wavelet-based SVC (WSVC). In the spatial decomposition module of DMSVC, the nonuniform directional distribution along scale of each frame is detected by phase congruency in the overcomplete wavelet domain. The corresponding orientational multi-resolution is achieved by nonuniform directional filter banks (NUDFB) which is fulfilled via a non-symmetric binary tree (NSBT) structured frequency division. The wavelet basis function in each scale is converted to an adaptive set of nonuniform directional basis by employing nonuniform directional filter banks. Experimental results validate a superior coding performance and visual quality over WSVC especially on those sequences full of directional edges and textures.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128876624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702471
Yongzhe Wang, N. Stefanoski, Xiangzhong Fang, A. Smolic
This paper presents an enhancement of the SVC extension of the H.264/AVC standard by content-adaptive spatial scalability (CASS). CASS introduces a novel functionality which is important for high quality content distribution. The video streams (spatial layers), which are used as input to the encoder, are created by content-adaptive and art-directable retargeting of existing high resolution video. Video is retargeted to resolutions and aspect ratios which are mainly dictated by target display devices. Thereby no content is cut off, but visually important content is preserved at the expense of a non-linear distortion of visually unimportant areas. The non-linear dependencies between such video streams are efficiently exploited by CASS for scalable coding. This is achieved by integrating warping-based non-linear texture prediction and warp coding into the SVC framework. The results indicate high prediction accuracy of non-linear predictors and high compression efficiency with limited increase in bit rate and complexity compared to the standard SVC for the case of INTRA only coding.
{"title":"Content-adaptive spatial scalability for scalable video coding","authors":"Yongzhe Wang, N. Stefanoski, Xiangzhong Fang, A. Smolic","doi":"10.1109/PCS.2010.5702471","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702471","url":null,"abstract":"This paper presents an enhancement of the SVC extension of the H.264/AVC standard by content-adaptive spatial scalability (CASS). CASS introduces a novel functionality which is important for high quality content distribution. The video streams (spatial layers), which are used as input to the encoder, are created by content-adaptive and art-directable retargeting of existing high resolution video. Video is retargeted to resolutions and aspect ratios which are mainly dictated by target display devices. Thereby no content is cut off, but visually important content is preserved at the expense of a non-linear distortion of visually unimportant areas. The non-linear dependencies between such video streams are efficiently exploited by CASS for scalable coding. This is achieved by integrating warping-based non-linear texture prediction and warp coding into the SVC framework. The results indicate high prediction accuracy of non-linear predictors and high compression efficiency with limited increase in bit rate and complexity compared to the standard SVC for the case of INTRA only coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127827566","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702547
Minghui Wang, Xin Jin, S. Goto
Depth map coding is a new topic in multiview video coding (MVC) following the development of depth-image-based rendering (DIBR). Since depth map is monochromatic and has less texture than color map, fast algorithm is necessary and possible to reduce the computation burden of the encoder. This paper proposed difference detection based early mode termination strategy. The difference detection (DD) algorithms are categorized to reconstructed frame based (RDD) and original frame based (ODD). A simplified ODD (sODD) strategy is also proposed. Early mode termination based on these three DD algorithms are implemented and evaluated in the reference software of Joint Multiview Video Coding (JMVC) version 8.0 respectively. Simulation results indicate that RDD based one has no performance lost and reduce 25% runtime on average. ODD and sODD based ones can save 54.3% and 43.6% runtime respectively and have an acceptable R-D performance lost.
{"title":"Difference detection based early mode termination for depth map coding in MVC","authors":"Minghui Wang, Xin Jin, S. Goto","doi":"10.1109/PCS.2010.5702547","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702547","url":null,"abstract":"Depth map coding is a new topic in multiview video coding (MVC) following the development of depth-image-based rendering (DIBR). Since depth map is monochromatic and has less texture than color map, fast algorithm is necessary and possible to reduce the computation burden of the encoder. This paper proposed difference detection based early mode termination strategy. The difference detection (DD) algorithms are categorized to reconstructed frame based (RDD) and original frame based (ODD). A simplified ODD (sODD) strategy is also proposed. Early mode termination based on these three DD algorithms are implemented and evaluated in the reference software of Joint Multiview Video Coding (JMVC) version 8.0 respectively. Simulation results indicate that RDD based one has no performance lost and reduce 25% runtime on average. ODD and sODD based ones can save 54.3% and 43.6% runtime respectively and have an acceptable R-D performance lost.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"166 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121538983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702491
Gene Cheung, Akira Kubota, Antonio Ortega
Compression of depth maps is important for “image plus depth” representation of multiview images, which enables synthesis of novel intermediate views via depth-image-based rendering (DIBR) at decoder. Previous depth map coding schemes exploit unique depth characteristics to compactly and faithfully reproduce the original signal. In contrast, given that depth maps are not directly viewed but are only used for view synthesis, in this paper we manipulate depth values themselves, without causing severe synthesized view distortion, in order to maximize sparsity in the transform domain for compression gain. We formulate the sparsity maximization problem as an l0-norm optimization. Given l0-norm optimization is hard in general, we first find a sparse representation by iteratively solving a weighted l1 minimization via linear programming (LP). We then design a heuristic to push resulting LP solution away from constraint boundaries to avoid quantization errors. Using JPEG as an example transform codec, we show that our approach gained up to 2.5dB in rate-distortion performance for the interpolated view.
{"title":"Sparse representation of depth maps for efficient transform coding","authors":"Gene Cheung, Akira Kubota, Antonio Ortega","doi":"10.1109/PCS.2010.5702491","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702491","url":null,"abstract":"Compression of depth maps is important for “image plus depth” representation of multiview images, which enables synthesis of novel intermediate views via depth-image-based rendering (DIBR) at decoder. Previous depth map coding schemes exploit unique depth characteristics to compactly and faithfully reproduce the original signal. In contrast, given that depth maps are not directly viewed but are only used for view synthesis, in this paper we manipulate depth values themselves, without causing severe synthesized view distortion, in order to maximize sparsity in the transform domain for compression gain. We formulate the sparsity maximization problem as an l0-norm optimization. Given l0-norm optimization is hard in general, we first find a sparse representation by iteratively solving a weighted l1 minimization via linear programming (LP). We then design a heuristic to push resulting LP solution away from constraint boundaries to avoid quantization errors. Using JPEG as an example transform codec, we show that our approach gained up to 2.5dB in rate-distortion performance for the interpolated view.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121558514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-12-01DOI: 10.1109/PCS.2010.5702563
P. Correia, P. Assunção, V. Silva
This paper proposes a multiple description (MD) transcoding scheme capable of preventing drift by distortion accumulation in temporally predicted motion compensated slices. Drift compensation is achieved by generating a controlled amount of side information to be used for decoding whenever a description fails to reach the end user terminal. The side information is generated by re-encoding the transcoding residue with an independent quantisation parameter which also controls redundancy. A simplified architecture is devised to reduce transcoding complexity in regard to the number of processing functions and buffer requirements. The experimental results show that temporally predicted frames do not suffer from drift and their quality is significantly improved at reduced redundancy cost in comparison with a classic MD transcoding scheme.
{"title":"Multiple description video transcoding with temporal drift control","authors":"P. Correia, P. Assunção, V. Silva","doi":"10.1109/PCS.2010.5702563","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702563","url":null,"abstract":"This paper proposes a multiple description (MD) transcoding scheme capable of preventing drift by distortion accumulation in temporally predicted motion compensated slices. Drift compensation is achieved by generating a controlled amount of side information to be used for decoding whenever a description fails to reach the end user terminal. The side information is generated by re-encoding the transcoding residue with an independent quantisation parameter which also controls redundancy. A simplified architecture is devised to reduce transcoding complexity in regard to the number of processing functions and buffer requirements. The experimental results show that temporally predicted frames do not suffer from drift and their quality is significantly improved at reduced redundancy cost in comparison with a classic MD transcoding scheme.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116408155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}