Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611943
Yunseok Song, Yo-Sung Ho
In this paper, we present a method to reduce complexity of depth modeling modes (DMM), which is currently used in the 3D-HEVC standardization activity. DMM adds four modes to the existing HEVC intra prediction modes; the main purpose is to accurately represent object edges in depth video. Mode 3 of DMM requires distortion calculation of all pre-defined wedgelets. The proposed method employs absolute differences of neighboring pixels in the reference block. The number of wedgelets that need to be concerned can be reduced to six. Experimental results show 3.1% complexity reduction on average while maintaining coding performance, which implies that the correct wedgelet is included, while non-viable wedgelets are disregarded.
{"title":"Simplified inter-component depth modeling in 3D-HEVC","authors":"Yunseok Song, Yo-Sung Ho","doi":"10.1109/IVMSPW.2013.6611943","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611943","url":null,"abstract":"In this paper, we present a method to reduce complexity of depth modeling modes (DMM), which is currently used in the 3D-HEVC standardization activity. DMM adds four modes to the existing HEVC intra prediction modes; the main purpose is to accurately represent object edges in depth video. Mode 3 of DMM requires distortion calculation of all pre-defined wedgelets. The proposed method employs absolute differences of neighboring pixels in the reference block. The number of wedgelets that need to be concerned can be reduced to six. Experimental results show 3.1% complexity reduction on average while maintaining coding performance, which implies that the correct wedgelet is included, while non-viable wedgelets are disregarded.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133059999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611917
Jing Li, M. Barkowsky, P. Callet
The measurement of the Quality of Experience (QoE) in 3DTV recently became an important research topic as it relates to the development of the 3D industry. Pair comparison is a reliable method as it is easier for the observers to provide their preference on a pair rather than give an absolute scale value to a stimulus. The QoE measured by pair comparison is thus called “Preference of Experience (PoE)”. In this paper, we introduce some efficient designs for pair comparison which can reduce the number of comparisons. The constraints of the presentation order of the stimuli in pair comparison test are listed. Finally, some analysis methods for pair comparison data are provided accompanied with some examples from the studies of the measurement of PoE.
{"title":"Subjective assessment methodology for preference of experience in 3DTV","authors":"Jing Li, M. Barkowsky, P. Callet","doi":"10.1109/IVMSPW.2013.6611917","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611917","url":null,"abstract":"The measurement of the Quality of Experience (QoE) in 3DTV recently became an important research topic as it relates to the development of the 3D industry. Pair comparison is a reliable method as it is easier for the observers to provide their preference on a pair rather than give an absolute scale value to a stimulus. The QoE measured by pair comparison is thus called “Preference of Experience (PoE)”. In this paper, we introduce some efficient designs for pair comparison which can reduce the number of comparisons. The constraints of the presentation order of the stimuli in pair comparison test are listed. Finally, some analysis methods for pair comparison data are provided accompanied with some examples from the studies of the measurement of PoE.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"82 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115029401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611899
Lee-Kang Liu, Zucheul Lee, Truong Nguyen
Recently, the work on dense disparity map reconstruction from 5% sparse initial estimates containing edges in disparity, has been proposed [1]. Practically, however, edges in disparity is unknown unless a dense disparity map has already been generated. In this paper, we present a realistic reconstruction framework for obtaining sharp and dense disparity maps from fixed number of sparse initial estimates with the aid of color image information. Experimental results show that sharp and dense disparity maps can be reconstructed at the cost of one pixel accuracy.
{"title":"Sharp disparity reconstruction using sparse disparity measurement and color information","authors":"Lee-Kang Liu, Zucheul Lee, Truong Nguyen","doi":"10.1109/IVMSPW.2013.6611899","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611899","url":null,"abstract":"Recently, the work on dense disparity map reconstruction from 5% sparse initial estimates containing edges in disparity, has been proposed [1]. Practically, however, edges in disparity is unknown unless a dense disparity map has already been generated. In this paper, we present a realistic reconstruction framework for obtaining sharp and dense disparity maps from fixed number of sparse initial estimates with the aid of color image information. Experimental results show that sharp and dense disparity maps can be reconstructed at the cost of one pixel accuracy.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"240 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123258658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611925
Woong Lim, I. Bajić, D. Sim
Rate control is an important component of an end-to-end video communication system. Although rate control is not a part of a video coding standard, it is necessary for practical deployment. Currently, there are several proposals for rate control in the upcoming High Efficiency Video Coding (HEVC) standard, but there is no rate control scheme for HEVC-based multi-view extension. In this paper, we apply the newly recommended R-λ model-based HEVC rate control to the multi-view scenario, and propose two improvements. One improvement deals with Quantization Parameter (QP) initialization, and the other deals with adaptive Mean Absolute Difference (MAD) prediction. Results demonstrate the accuracy of the proposed methods, the resulting reduced fluctuation of instantaneous bitrate, as well as an improvement in rate-distortion performance compared to the R-λ rate control alone.
{"title":"QP initialization and adaptive MAD prediction for rate control in HEVC-based multi-view video coding","authors":"Woong Lim, I. Bajić, D. Sim","doi":"10.1109/IVMSPW.2013.6611925","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611925","url":null,"abstract":"Rate control is an important component of an end-to-end video communication system. Although rate control is not a part of a video coding standard, it is necessary for practical deployment. Currently, there are several proposals for rate control in the upcoming High Efficiency Video Coding (HEVC) standard, but there is no rate control scheme for HEVC-based multi-view extension. In this paper, we apply the newly recommended R-λ model-based HEVC rate control to the multi-view scenario, and propose two improvements. One improvement deals with Quantization Parameter (QP) initialization, and the other deals with adaptive Mean Absolute Difference (MAD) prediction. Results demonstrate the accuracy of the proposed methods, the resulting reduced fluctuation of instantaneous bitrate, as well as an improvement in rate-distortion performance compared to the R-λ rate control alone.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123174767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611916
Philippe Hanhart, T. Ebrahimi
Development of objective quality metrics that can reliably predict perceived quality of 3D video sequences is challenging. Various 3D objective metrics have been proposed, but PSNR is still widely used. Several studies have shown that PSNR is strongly content dependent, but the exact relationship between PSNR values and perceived quality has not been established yet. In this paper, we propose a model to predict the relationship between PSNR values and perceived quality of stereoscopic video sequences based on content analysis. The model was trained and evaluated on a dataset of stereoscopic video sequences with associated ground truth MOS. Results showed that the proposed model achieved high correlation with perceived quality and was quite robust across contents when the training set contained various contents.
{"title":"Predicting 3D quality based on content analysis","authors":"Philippe Hanhart, T. Ebrahimi","doi":"10.1109/IVMSPW.2013.6611916","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611916","url":null,"abstract":"Development of objective quality metrics that can reliably predict perceived quality of 3D video sequences is challenging. Various 3D objective metrics have been proposed, but PSNR is still widely used. Several studies have shown that PSNR is strongly content dependent, but the exact relationship between PSNR values and perceived quality has not been established yet. In this paper, we propose a model to predict the relationship between PSNR values and perceived quality of stereoscopic video sequences based on content analysis. The model was trained and evaluated on a dataset of stereoscopic video sequences with associated ground truth MOS. Results showed that the proposed model achieved high correlation with perceived quality and was quite robust across contents when the training set contained various contents.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133495655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611922
Kazuki Matsumoto, Chiyoung Song, François de Sorbier, H. Saito
In this paper, we propose an approach for generating free viewpoint videos based on multiple depth and color cameras to resolve issues encountered with traditional color cameras techniques. Our system is based on consumer products such as Kinect that does not provide satisfying quality in terms of resolution and noise. Our contribution is then to propose a full pipeline for enhancing the depth maps and finally improving the quality of the novel viewpoint generated.
{"title":"Free viewpoint video synthesis using multi-view depth and color cameras","authors":"Kazuki Matsumoto, Chiyoung Song, François de Sorbier, H. Saito","doi":"10.1109/IVMSPW.2013.6611922","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611922","url":null,"abstract":"In this paper, we propose an approach for generating free viewpoint videos based on multiple depth and color cameras to resolve issues encountered with traditional color cameras techniques. Our system is based on consumer products such as Kinect that does not provide satisfying quality in terms of resolution and noise. Our contribution is then to propose a full pipeline for enhancing the depth maps and finally improving the quality of the novel viewpoint generated.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130113405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611942
Wei Yong Eng, Dongbo Min, V. Nguyen, Jiangbo Lu, M. Do
The lack of eye contact between participants in a tele-conferencing makes nonverbal communication unnatural and ineffective. A lot of research has focused on correcting the user gaze for a natural communication. Most of prior solutions require expensive and bulky hardware, or incorporate a complicated algorithm causing inefficiency and deployment. In this paper, we propose an effective and efficient gaze correction solution for a 3D tele-conferencing system in a single color/depth camera set-up. A raw depth map is first refined using the corresponding color image. Then, both color and depth data of the participant are accurately segmented. A novel view is synthesized in the location of the display screen which coincides with the user gaze. Stereoscopic views, i.e. virtual left and right images, can also be generated for 3D immersive conferencing, and are displayed in a 3D monitor with 3D virtual background scenes. Finally, to handle large hole regions that often occur in the view synthesized with a single color camera, we propose a simple yet robust hole filling technique that works in real-time. This novel inpainting method can effectively reconstruct missing parts of the synthesized image under various challenging situations. Our proposed system works in real-time on a single core CPU without requiring dedicated hardware, including data acquisition, post-processing, rendering, and so on.
{"title":"Gaze correction for 3D tele-immersive communication system","authors":"Wei Yong Eng, Dongbo Min, V. Nguyen, Jiangbo Lu, M. Do","doi":"10.1109/IVMSPW.2013.6611942","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611942","url":null,"abstract":"The lack of eye contact between participants in a tele-conferencing makes nonverbal communication unnatural and ineffective. A lot of research has focused on correcting the user gaze for a natural communication. Most of prior solutions require expensive and bulky hardware, or incorporate a complicated algorithm causing inefficiency and deployment. In this paper, we propose an effective and efficient gaze correction solution for a 3D tele-conferencing system in a single color/depth camera set-up. A raw depth map is first refined using the corresponding color image. Then, both color and depth data of the participant are accurately segmented. A novel view is synthesized in the location of the display screen which coincides with the user gaze. Stereoscopic views, i.e. virtual left and right images, can also be generated for 3D immersive conferencing, and are displayed in a 3D monitor with 3D virtual background scenes. Finally, to handle large hole regions that often occur in the view synthesized with a single color camera, we propose a simple yet robust hole filling technique that works in real-time. This novel inpainting method can effectively reconstruct missing parts of the synthesized image under various challenging situations. Our proposed system works in real-time on a single core CPU without requiring dedicated hardware, including data acquisition, post-processing, rendering, and so on.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130348121","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611907
Sotirios Delis, N. Nikolaidis, I. Pitas
3DTV and 3D cinema witness a significant increase in their popularity nowadays. New movie titles are released in 3D and there are more than 35 TV channels in various countries that broadcast in 3D worldwide. It is well known today and becomes more obvious, as the 3D video content availability increases, that stereoscopy is associated with certain 3D video quality issues that may affect in a negative way the 3D viewing experience. In this paper, we propose two novel algorithms that exploit available disparity information to detect two disturbing stereoscopic issues, namely depth jump cuts and bent window effects. Representative examples are provided to assess the algorithms performance. The proposed algorithms can be helpful in the post-production stage, where, in most cases, the detected issues can be fixed, and also in assessing the overall quality of stereoscopic video content.
{"title":"Automatic detection of depth jump cuts and bent window effects in stereoscopic videos","authors":"Sotirios Delis, N. Nikolaidis, I. Pitas","doi":"10.1109/IVMSPW.2013.6611907","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611907","url":null,"abstract":"3DTV and 3D cinema witness a significant increase in their popularity nowadays. New movie titles are released in 3D and there are more than 35 TV channels in various countries that broadcast in 3D worldwide. It is well known today and becomes more obvious, as the 3D video content availability increases, that stereoscopy is associated with certain 3D video quality issues that may affect in a negative way the 3D viewing experience. In this paper, we propose two novel algorithms that exploit available disparity information to detect two disturbing stereoscopic issues, namely depth jump cuts and bent window effects. Representative examples are provided to assess the algorithms performance. The proposed algorithms can be helpful in the post-production stage, where, in most cases, the detected issues can be fixed, and also in assessing the overall quality of stereoscopic video content.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"515 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123081970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611927
Seungha Yang, Hoyoung Lee, H. Shim, B. Jeon
In this paper, we propose a fast inter mode decision method for High Efficiency Video Coding (HEVC) in order to reduce computational complexity of its encoder. It utilizes correlated tendency of PU mode. Compared to the early termination schemes for fast mode decision already implemented in HEVC reference software, it reduces the loss of coding efficiency. Experimental results show that the proposed method decreases encoding time by about 22.5% with little increment in bit-rate. Furthermore, the proposed method shows encoding time reduction of 39.1% with only 0.8% increase of bit-rate when it is combined with existing fast methods such as early CU termination scheme.
{"title":"Fast inter mode decision process for HEVC encoder","authors":"Seungha Yang, Hoyoung Lee, H. Shim, B. Jeon","doi":"10.1109/IVMSPW.2013.6611927","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611927","url":null,"abstract":"In this paper, we propose a fast inter mode decision method for High Efficiency Video Coding (HEVC) in order to reduce computational complexity of its encoder. It utilizes correlated tendency of PU mode. Compared to the early termination schemes for fast mode decision already implemented in HEVC reference software, it reduces the loss of coding efficiency. Experimental results show that the proposed method decreases encoding time by about 22.5% with little increment in bit-rate. Furthermore, the proposed method shows encoding time reduction of 39.1% with only 0.8% increase of bit-rate when it is combined with existing fast methods such as early CU termination scheme.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"117 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115703257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611892
Dae-Sik Kim, Ho-Sup Lee, S. Shestak, SungWoo Cho
Ambient light inside viewing field of shutter-glasses 3DTV system can cause perceivable flicker due to high brightness of the light source. Omitting front polarizer of shutter glasses can be a solution for improving ambient light flicker, but it makes noticeable ghosting whenever 3D viewers tilt their heads. In this paper, we propose the new flicker-free shutter glasses compensated for viewers head tilt using tilt sensor. The crosstalk level, inserted by the shutter is below 1.6% within the tilt angle range from 0 to ±50°.
{"title":"Flicker-free 3D shutter glasses by retardnace control of LC cell","authors":"Dae-Sik Kim, Ho-Sup Lee, S. Shestak, SungWoo Cho","doi":"10.1109/IVMSPW.2013.6611892","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611892","url":null,"abstract":"Ambient light inside viewing field of shutter-glasses 3DTV system can cause perceivable flicker due to high brightness of the light source. Omitting front polarizer of shutter glasses can be a solution for improving ambient light flicker, but it makes noticeable ghosting whenever 3D viewers tilt their heads. In this paper, we propose the new flicker-free shutter glasses compensated for viewers head tilt using tilt sensor. The crosstalk level, inserted by the shutter is below 1.6% within the tilt angle range from 0 to ±50°.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114567765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}