Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611906
Kwanghyun Lee, Haksub Kim, Sanghoon Lee
One of the most challenging issues in the 3D visual research field is how to quantify the visualization displayed over the virtual 3D space. To seek an effective method of quantification, it is necessary to measure various important elements related to different depths of 3D objects. In this paper, we propose a new framework to quantify the 3D visual information, termed 3D activity by measuring natural scene statistics (NSS). In the simulation, we verify the effectiveness of 3D activity to quantify the degree of freedom of 3D space in various aspects: disparity and motion.
{"title":"3D activity measurement for stereoscopic video","authors":"Kwanghyun Lee, Haksub Kim, Sanghoon Lee","doi":"10.1109/IVMSPW.2013.6611906","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611906","url":null,"abstract":"One of the most challenging issues in the 3D visual research field is how to quantify the visualization displayed over the virtual 3D space. To seek an effective method of quantification, it is necessary to measure various important elements related to different depths of 3D objects. In this paper, we propose a new framework to quantify the 3D visual information, termed 3D activity by measuring natural scene statistics (NSS). In the simulation, we verify the effectiveness of 3D activity to quantify the degree of freedom of 3D space in various aspects: disparity and motion.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114547682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611929
M. Grum, A. Bors
In this paper we present a new approach for modeling and correcting scenes with multiple 3D objects from images taken from various viewpoints. For the 3D scene initialization we consider implicit radial basis functions (RBF) estimated from the voxel model produced by the space carving algorithm. 3D scenes are corrected using image content disparities within its image projections as well as its inconsistency with its silhouettes extracted from images. While the image content disparities are suitable for textured regions, the silhouettes can be applied to regions of uniform colour which can be accurately segmented.
{"title":"3D scene correction using disparities with its projections","authors":"M. Grum, A. Bors","doi":"10.1109/IVMSPW.2013.6611929","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611929","url":null,"abstract":"In this paper we present a new approach for modeling and correcting scenes with multiple 3D objects from images taken from various viewpoints. For the 3D scene initialization we consider implicit radial basis functions (RBF) estimated from the voxel model produced by the space carving algorithm. 3D scenes are corrected using image content disparities within its image projections as well as its inconsistency with its silhouettes extracted from images. While the image content disparities are suitable for textured regions, the silhouettes can be applied to regions of uniform colour which can be accurately segmented.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124231014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611941
O. Schreer, M. Bertzen, N. Atzpadin, C. Riechert, W. Waizenegger, I. Feldmann
Multi-view camera calibration is an essential task in the filed of 3D reconstruction which holds especially for immersive media applications like 3D videocommunication. Although the problem of multi-view calibration is basically solved, there is still space to improve the calibration process and to increase the accuracy during acquisition of calibration patterns. It is commonly known that robust and accurate calibration requires feature points that are equally distributed in 3D space covering the whole volume of interest. In this paper, we propose a user guided calibration based on a graphical user interface, which drastically simplifies the correct acquisition of calibration patterns. Based on an optimized selection of patterns and their corresponding feature points, the multi-view calibration becomes much faster in terms of data acquisition as well as computational effort by reaching the same accuracy with standard unguided acquisitions of calibration pattern.
{"title":"User-feedback and optimization for multi-view calibration","authors":"O. Schreer, M. Bertzen, N. Atzpadin, C. Riechert, W. Waizenegger, I. Feldmann","doi":"10.1109/IVMSPW.2013.6611941","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611941","url":null,"abstract":"Multi-view camera calibration is an essential task in the filed of 3D reconstruction which holds especially for immersive media applications like 3D videocommunication. Although the problem of multi-view calibration is basically solved, there is still space to improve the calibration process and to increase the accuracy during acquisition of calibration patterns. It is commonly known that robust and accurate calibration requires feature points that are equally distributed in 3D space covering the whole volume of interest. In this paper, we propose a user guided calibration based on a graphical user interface, which drastically simplifies the correct acquisition of calibration patterns. Based on an optimized selection of patterns and their corresponding feature points, the multi-view calibration becomes much faster in terms of data acquisition as well as computational effort by reaching the same accuracy with standard unguided acquisitions of calibration pattern.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"94 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126221256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611920
J. Heo
We provide an important analysis of depth variation of human faces. Throughout an extensive analysis of 3D face shapes, we claim that 3D depth information (z) of faces is not significantly changing and can be synthesized from another person's depth or a generic depth information. We also show that gender and ethnicity specific average depth models can approximate the 3D shape of the input face image more accurately, achieving a better generalization of 3D face modeling and reconstruction compared to a global average depth model.
{"title":"3D depth analysis of human faces","authors":"J. Heo","doi":"10.1109/IVMSPW.2013.6611920","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611920","url":null,"abstract":"We provide an important analysis of depth variation of human faces. Throughout an extensive analysis of 3D face shapes, we claim that 3D depth information (z) of faces is not significantly changing and can be synthesized from another person's depth or a generic depth information. We also show that gender and ethnicity specific average depth models can approximate the 3D shape of the input face image more accurately, achieving a better generalization of 3D face modeling and reconstruction compared to a global average depth model.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130466204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611912
Ji-Ho Cho, Satoshi Ikehata, H. Yoo, M. Gelautz, K. Aizawa
Depth maps captured by active sensors (e.g., ToF cameras and Kinect) typically suffer from poor spatial resolution, considerable amount of noise, and missing data. To overcome these problems, we propose a novel depth map up-sampling method which increases the resolution of the original depth map while effectively suppressing aliasing artifacts. Assuming that a registered high-resolution texture image is available, the cost-volume filtering framework is applied to this problem. Our experiments show that cost-volume filtering can generate the high-resolution depth map accurately and efficiently while preserving discontinuous object boundaries, which is often a challenge when various state-of-the-art algorithms are applied.
{"title":"Depth map up-sampling using cost-volume filtering","authors":"Ji-Ho Cho, Satoshi Ikehata, H. Yoo, M. Gelautz, K. Aizawa","doi":"10.1109/IVMSPW.2013.6611912","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611912","url":null,"abstract":"Depth maps captured by active sensors (e.g., ToF cameras and Kinect) typically suffer from poor spatial resolution, considerable amount of noise, and missing data. To overcome these problems, we propose a novel depth map up-sampling method which increases the resolution of the original depth map while effectively suppressing aliasing artifacts. Assuming that a registered high-resolution texture image is available, the cost-volume filtering framework is applied to this problem. Our experiments show that cost-volume filtering can generate the high-resolution depth map accurately and efficiently while preserving discontinuous object boundaries, which is often a challenge when various state-of-the-art algorithms are applied.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130825691","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611924
Jung-Hak Nam, I. Bajić, D. Sim
In this paper, we present an efficient motion and disparity prediction method for multi-view video coding based on the high efficient video coding (HEVC) standard. The proposed method exploits inter-view candidates for effective prediction of the motion or disparity vector to be coded. The inter-view candidates include not only motion vectors of adjacent views, but also global disparities across views. We found that motion vectors coded earlier in an adjacent view are helpful in predicting the current motion vector to reduce the amount of bits used in the motion vector information. In addition, the proposed disparity prediction with global disparity method is effective for inter-view prediction. To evaluate the proposed algorithm, we implemented a multi-view version based on HEVC and the proposed correspondence prediction method is implemented on a multi-view platform based on HEVC. We found that the proposed algorithm yields a coding gain of around 2.9% in a high efficiency configuration random access mode.
{"title":"Novel motion prediction for multi-view video coding using global disparity","authors":"Jung-Hak Nam, I. Bajić, D. Sim","doi":"10.1109/IVMSPW.2013.6611924","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611924","url":null,"abstract":"In this paper, we present an efficient motion and disparity prediction method for multi-view video coding based on the high efficient video coding (HEVC) standard. The proposed method exploits inter-view candidates for effective prediction of the motion or disparity vector to be coded. The inter-view candidates include not only motion vectors of adjacent views, but also global disparities across views. We found that motion vectors coded earlier in an adjacent view are helpful in predicting the current motion vector to reduce the amount of bits used in the motion vector information. In addition, the proposed disparity prediction with global disparity method is effective for inter-view prediction. To evaluate the proposed algorithm, we implemented a multi-view version based on HEVC and the proposed correspondence prediction method is implemented on a multi-view platform based on HEVC. We found that the proposed algorithm yields a coding gain of around 2.9% in a high efficiency configuration random access mode.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127867538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611908
Hosik Sohn, Yong Ju Jung, Seong-il Lee, Yong Man Ro
This paper proposes a novel crosstalk reduction method for stereoscopic 3D displays by using a combined approach of disparity adjustment and subtractive crosstalk cancellation. Specifically, we propose a disparity adjustment method that can minimize the perceived crosstalk and negative effects of the crosstalk cancellation on the image quality. In addition, we also provide a contrast reduction method optimized for the subtractive crosstalk cancellation. The experimental results showed that the proposed method could provide higher image quality than existing crosstalk cancellation methods while successfully reducing the perceived crosstalk.
{"title":"Crosstalk reduction in stereoscopic displays: A combined approach of disparity adjustment and crosstalk cancellation","authors":"Hosik Sohn, Yong Ju Jung, Seong-il Lee, Yong Man Ro","doi":"10.1109/IVMSPW.2013.6611908","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611908","url":null,"abstract":"This paper proposes a novel crosstalk reduction method for stereoscopic 3D displays by using a combined approach of disparity adjustment and subtractive crosstalk cancellation. Specifically, we propose a disparity adjustment method that can minimize the perceived crosstalk and negative effects of the crosstalk cancellation on the image quality. In addition, we also provide a contrast reduction method optimized for the subtractive crosstalk cancellation. The experimental results showed that the proposed method could provide higher image quality than existing crosstalk cancellation methods while successfully reducing the perceived crosstalk.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133725113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611923
Hansung Kim, A. Hilton
We propose a plane-based urban scene reconstruction method using spherical stereo image pairs. We assume that the urban scene consists of axis-aligned approximately planar structures (Manhattan world). Captured spherical stereo images are converted into six central-point perspective images by cubic projection and facade alignment. Facade alignment automatically identifies the principal planes direction in the scene allowing the cubic projection to preserve the plane structure. Depth information is recovered by stereo matching between images and independent 3D rectangular planes are constructed by plane fitting aligned with the principal axes. Finally planar regions are refined by expanding, detecting intersections and cropping based on visibility. The reconstructed model efficiently represents the structure of the scene and texture mapping allows natural walk-through rendering.
{"title":"Planar urban scene reconstruction from spherical images using facade alignment","authors":"Hansung Kim, A. Hilton","doi":"10.1109/IVMSPW.2013.6611923","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611923","url":null,"abstract":"We propose a plane-based urban scene reconstruction method using spherical stereo image pairs. We assume that the urban scene consists of axis-aligned approximately planar structures (Manhattan world). Captured spherical stereo images are converted into six central-point perspective images by cubic projection and facade alignment. Facade alignment automatically identifies the principal planes direction in the scene allowing the cubic projection to preserve the plane structure. Depth information is recovered by stereo matching between images and independent 3D rectangular planes are constructed by plane fitting aligned with the principal axes. Finally planar regions are refined by expanding, detecting intersections and cropping based on visibility. The reconstructed model efficiently represents the structure of the scene and texture mapping allows natural walk-through rendering.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128473271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611935
M. Ko, Dongwook Kim, Jisang Yoo
In this paper, a new common-hole filling algorithm with a probability mask for a virtual view synthesis is proposed. In the proposed algorithm, we try to combine strong points of both the spiral weighted average algorithm and the gradient searching algorithm. The spiral weighted average algorithm keeps the boundary of each object well by using depth information and the gradient searching algorithm is able to preserve details. We also try to reduce the flickering defect existing around the filled common-hole region by using a probability mask. The experimental results show that the proposed algorithm performs much better than conventional algorithms.
{"title":"A new common-hole filling algorithm for virtual view synthesis with a probability mask","authors":"M. Ko, Dongwook Kim, Jisang Yoo","doi":"10.1109/IVMSPW.2013.6611935","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611935","url":null,"abstract":"In this paper, a new common-hole filling algorithm with a probability mask for a virtual view synthesis is proposed. In the proposed algorithm, we try to combine strong points of both the spiral weighted average algorithm and the gradient searching algorithm. The spiral weighted average algorithm keeps the boundary of each object well by using depth information and the gradient searching algorithm is able to preserve details. We also try to reduce the flickering defect existing around the filled common-hole region by using a probability mask. The experimental results show that the proposed algorithm performs much better than conventional algorithms.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"118 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132711722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-06-10DOI: 10.1109/IVMSPW.2013.6611915
Bochao Zou, Yue Liu, Yongtian Wang, Tao Huang, Q. Zhu
Prior researches on binocular mismatches of stereoscopic images mostly focus on optical errors (magnification, shift, rotation, distortion) and photometric asymmetries (color, luminance, definition). In this paper, a type of binocular mismatches - effect of absence is investigated to find whether the partial loss of an object in either one of a stereo pair influences characteristics of human fusion and depth perception when certain disparity is provided, in which two situations are taken into account: with overlap and without overlap. Implications for the effect of absence on fusion mechanism and visual comfort are also discussed in this paper. Experimental results prove the conclusion that the effect of absence can cause misperception and contribute to visual discomfort.
{"title":"Effect of absence on visual perception and discomfort","authors":"Bochao Zou, Yue Liu, Yongtian Wang, Tao Huang, Q. Zhu","doi":"10.1109/IVMSPW.2013.6611915","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611915","url":null,"abstract":"Prior researches on binocular mismatches of stereoscopic images mostly focus on optical errors (magnification, shift, rotation, distortion) and photometric asymmetries (color, luminance, definition). In this paper, a type of binocular mismatches - effect of absence is investigated to find whether the partial loss of an object in either one of a stereo pair influences characteristics of human fusion and depth perception when certain disparity is provided, in which two situations are taken into account: with overlap and without overlap. Implications for the effect of absence on fusion mechanism and visual comfort are also discussed in this paper. Experimental results prove the conclusion that the effect of absence can cause misperception and contribute to visual discomfort.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124286399","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}