Pub Date : 2011-12-05DOI: 10.1109/IC3D.2011.6584376
Lasith Yasakethu, L. Blondé, D. Doyen, Q. Huynh-Thu
3D cinema and 3DTV are at two different levels in the screen size spectrum. When the same stereoscopic-3D content is viewed on a cinema screen and 3DTV screen, it will produce a different 3D impression. As a result, it is difficult to fulfill the requirements of 3DTV with content captured for 3D cinema. Thus, it is important to properly address the issue of 3DTV content creation to avoid possible delays in the deployment of 3DTV. In this paper, we first explore the effects of using the same content for 3D cinema and 3DTV and then analyze the performance of several disparity based transformations for 3D cinema to 3DTV content conversion, by subjective testing. Effectiveness of the transformations is analyzed in terms of both depth quality and visual comfort of 3D experience. We show that by using a simple shift-based disparity transformation technique, it is possible to enhance the 3DTV experience from a common input signal which is originally captured for cinema viewing.
{"title":"Transforming 3D cinema content for an enhanced 3DTV experience","authors":"Lasith Yasakethu, L. Blondé, D. Doyen, Q. Huynh-Thu","doi":"10.1109/IC3D.2011.6584376","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584376","url":null,"abstract":"3D cinema and 3DTV are at two different levels in the screen size spectrum. When the same stereoscopic-3D content is viewed on a cinema screen and 3DTV screen, it will produce a different 3D impression. As a result, it is difficult to fulfill the requirements of 3DTV with content captured for 3D cinema. Thus, it is important to properly address the issue of 3DTV content creation to avoid possible delays in the deployment of 3DTV. In this paper, we first explore the effects of using the same content for 3D cinema and 3DTV and then analyze the performance of several disparity based transformations for 3D cinema to 3DTV content conversion, by subjective testing. Effectiveness of the transformations is analyzed in terms of both depth quality and visual comfort of 3D experience. We show that by using a simple shift-based disparity transformation technique, it is possible to enhance the 3DTV experience from a common input signal which is originally captured for cinema viewing.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126246392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584372
Takuya Yoshii, Shu Matsuura
In this study, we created a three-dimensional (3D) model of a simplified direct current (DC) motor and used it to explain the mechanism of actual DC motors to junior high school science classes. We implemented a questionnaire for students before and after the presentation of the model. Before the presentation, many students were not confident about explaining the mechanism of a DC motor based on Fleming's left-hand rule. Then, we showed the stereoscopic display of our DC motor model from various viewpoints and explained the application of Fleming's left-hand rule. The results of the questionnaire suggested that students gained confidence in explaining the application after viewing the stereoscopic display. It was suggested that the change of viewpoint in the stereoscopic display was effective in improving their understanding.
{"title":"Educational benefits of stereoscopic visualization from multiple viewpoints, illustrated with an electrical motor model","authors":"Takuya Yoshii, Shu Matsuura","doi":"10.1109/IC3D.2011.6584372","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584372","url":null,"abstract":"In this study, we created a three-dimensional (3D) model of a simplified direct current (DC) motor and used it to explain the mechanism of actual DC motors to junior high school science classes. We implemented a questionnaire for students before and after the presentation of the model. Before the presentation, many students were not confident about explaining the mechanism of a DC motor based on Fleming's left-hand rule. Then, we showed the stereoscopic display of our DC motor model from various viewpoints and explained the application of Fleming's left-hand rule. The results of the questionnaire suggested that students gained confidence in explaining the application after viewing the stereoscopic display. It was suggested that the change of viewpoint in the stereoscopic display was effective in improving their understanding.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"81 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125888703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584377
Wauthier d'Ursel
Device for presenting 3D animated images, obtained by taking pictures through a diffraction grating, from all size objects, and a projection onto a holographic screen. The diffraction, orientated horizontally, gets by each ray of light a way done which depends on its wavelength and allows to obtain as many angles of vision as wavelengths. The screen diffracts horizontally and works as a complex lens reconstituing an integral 3D effect. Each pixel of the recording support realises a spectrum analysis and, at the time of the projection, return the information. With three recording supports, RGB natural colours are obtained. The data are converted into digital form. The data processing allows to create synthetic three-dimentional images.
{"title":"3D holographic video system","authors":"Wauthier d'Ursel","doi":"10.1109/IC3D.2011.6584377","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584377","url":null,"abstract":"Device for presenting 3D animated images, obtained by taking pictures through a diffraction grating, from all size objects, and a projection onto a holographic screen. The diffraction, orientated horizontally, gets by each ray of light a way done which depends on its wavelength and allows to obtain as many angles of vision as wavelengths. The screen diffracts horizontally and works as a complex lens reconstituing an integral 3D effect. Each pixel of the recording support realises a spectrum analysis and, at the time of the projection, return the information. With three recording supports, RGB natural colours are obtained. The data are converted into digital form. The data processing allows to create synthetic three-dimentional images.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129490982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584375
Yilong Liu, Yuanyuan Jiang, Yebin Liu
An image based multi-view reconstruction system, fast and robust, is introduced in this article. Instead of using visual hull as input, multi-view images are directly used to generate a precise and watertight 3D model. At the same time, point cloud and visual hull can also be produced as by-products. Our system is made up of three stages: point cloud generation, fusion and meshing. With well-designed algorithm and data structure, state-of-the-art speed is achieved in this highly reliable system, as revealed in our comparison with related work.
{"title":"A visual hull free algorithm for fast and robust multi-view stereo","authors":"Yilong Liu, Yuanyuan Jiang, Yebin Liu","doi":"10.1109/IC3D.2011.6584375","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584375","url":null,"abstract":"An image based multi-view reconstruction system, fast and robust, is introduced in this article. Instead of using visual hull as input, multi-view images are directly used to generate a precise and watertight 3D model. At the same time, point cloud and visual hull can also be produced as by-products. Our system is made up of three stages: point cloud generation, fusion and meshing. With well-designed algorithm and data structure, state-of-the-art speed is achieved in this highly reliable system, as revealed in our comparison with related work.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124468748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584371
Jordi Salvador, J. Casas
This paper presents a methodology for obtaining a 3D reconstruction of a dynamic scene in multi-camera settings. Our target is to derive a compact representation of the 3D scene which is effective and accurate, whatever the number of cameras and even for very-wide baseline settings. Easing real-time 3D scene capture has outstanding applications in 2D and 3D content production, free viewpoint video of natural scenes and interactive video applications.
{"title":"A compact 3D representation for multi-view video","authors":"Jordi Salvador, J. Casas","doi":"10.1109/IC3D.2011.6584371","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584371","url":null,"abstract":"This paper presents a methodology for obtaining a 3D reconstruction of a dynamic scene in multi-camera settings. Our target is to derive a compact representation of the 3D scene which is effective and accurate, whatever the number of cameras and even for very-wide baseline settings. Easing real-time 3D scene capture has outstanding applications in 2D and 3D content production, free viewpoint video of natural scenes and interactive video applications.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116809094","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584389
Karim Benzeroual, L. Wilcox, Ali Kazimi, R. Allison
A primary concern when making stereoscopic 3D (S3D) movies is to promote an effective and comfortable S3D experience for the audience when displayed on the screen. The amount of depth produced on-screen can be controlled using a variety of parameters. Many of these are lighting related such as lighting architecture and technology. Others are optical or positional and thus have a geometrical effect including camera interaxial distance, camera convergence, lens properties, viewing distance and angle, screen/projector properties and viewer anatomy (interocular distance). The amount of estimated depth from disparity alone can be precisely predicted from simple trigonometry; however, perceived depth from disparity in complex scenes is difficult to evaluate and most likely different from the predicted depth based on geometry. This discrepancy is mediated by perceptual and cognitive factors, including resolution of the combination/conflict of pictorial, motion and binocular depth cues. This paper will review geometric predictions of depth from disparity and present the results of experiments which assess perceived S3D depth and the effect of the complexity of scene content.
{"title":"On the distinction between perceived & predicted depth in S3D films","authors":"Karim Benzeroual, L. Wilcox, Ali Kazimi, R. Allison","doi":"10.1109/IC3D.2011.6584389","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584389","url":null,"abstract":"A primary concern when making stereoscopic 3D (S3D) movies is to promote an effective and comfortable S3D experience for the audience when displayed on the screen. The amount of depth produced on-screen can be controlled using a variety of parameters. Many of these are lighting related such as lighting architecture and technology. Others are optical or positional and thus have a geometrical effect including camera interaxial distance, camera convergence, lens properties, viewing distance and angle, screen/projector properties and viewer anatomy (interocular distance). The amount of estimated depth from disparity alone can be precisely predicted from simple trigonometry; however, perceived depth from disparity in complex scenes is difficult to evaluate and most likely different from the predicted depth based on geometry. This discrepancy is mediated by perceptual and cognitive factors, including resolution of the combination/conflict of pictorial, motion and binocular depth cues. This paper will review geometric predictions of depth from disparity and present the results of experiments which assess perceived S3D depth and the effect of the complexity of scene content.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128914258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584393
A. Lejeune, S. Piérard, Marc Van Droogenbroeck, J. Verly
Edges are a fundamental clue for analyzing, interpreting, and understanding 3D scenes: they describe objects boundaries. Available edge detection methods are not suited for 3D cameras such as the Microsoft Kinect or a time-of-flight camera: they are slow and do not take into consideration the characteristics of the cameras. In this paper, we present a fast jump edge detection technique for 3D cameras based on the principles of Canny's edge detector. We first analyze the characteristics of the range signal for two different kinds of cameras: a time-of-flight camera (the PMD[vision] CamCube) and the Microsoft Kinect. From this analysis, we define appropriate operators and thresholds to perform the edge detection. Then, we present some results of the developed algorithms for both cameras.
{"title":"A new jump edge detection method for 3D cameras","authors":"A. Lejeune, S. Piérard, Marc Van Droogenbroeck, J. Verly","doi":"10.1109/IC3D.2011.6584393","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584393","url":null,"abstract":"Edges are a fundamental clue for analyzing, interpreting, and understanding 3D scenes: they describe objects boundaries. Available edge detection methods are not suited for 3D cameras such as the Microsoft Kinect or a time-of-flight camera: they are slow and do not take into consideration the characteristics of the cameras. In this paper, we present a fast jump edge detection technique for 3D cameras based on the principles of Canny's edge detector. We first analyze the characteristics of the range signal for two different kinds of cameras: a time-of-flight camera (the PMD[vision] CamCube) and the Microsoft Kinect. From this analysis, we define appropriate operators and thresholds to perform the edge detection. Then, we present some results of the developed algorithms for both cameras.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116103986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584366
Kun Xu, Xiangyang Ji, Ruiping Wang, Qionghai Dai
Depth-image-based rendering (DIBR) is a key step in 3D video generation. Parallel implementation of DIBR is able to improve rendering efficiency. General DIBR algorithms include two steps: pixel shifting (warping) and hole filling. There are memory correlations in these steps. To minimize memory conflict, we employ an auxiliary matrix to record maximum shifting distance. Implementation details on OpenMP and CUDA are presented and experimental results on GPU and multi-core CPU are compared.
{"title":"Parallel implementation of depth-image-based rendering","authors":"Kun Xu, Xiangyang Ji, Ruiping Wang, Qionghai Dai","doi":"10.1109/IC3D.2011.6584366","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584366","url":null,"abstract":"Depth-image-based rendering (DIBR) is a key step in 3D video generation. Parallel implementation of DIBR is able to improve rendering efficiency. General DIBR algorithms include two steps: pixel shifting (warping) and hole filling. There are memory correlations in these steps. To minimize memory conflict, we employ an auxiliary matrix to record maximum shifting distance. Implementation details on OpenMP and CUDA are presented and experimental results on GPU and multi-core CPU are compared.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123868195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584364
Qian Ma, Yongbing Zhang, Qiong Liu, Xiangyang Ji, Qionghai Dai
Stereo interleaving format attracts more and more attention due to its back compatibility with all existing 2D video coding standards as well as its high efficiency of stereoscopic video compression. As one of the most significant coding components, rate-distortion optimization (RDO) in stereo interleaving video coding does not take into account its application scenarios where the reconstructed video needs up-sampling for displaying. To promote the efficiency of stereo interleaving video compression, we propose an enhanced RDO where the up-sampling is taken into consideration on distortion measurement. The proposed algorithm changes nothing concerning the syntax and is compatible with current decoders. Experimental results demonstrate that the proposed enhanced RDO is able to reduce the bitrates by 10%-44% on average, and achieve 0.1-0.65dB gain on PSNR compared with conventional RDO.
{"title":"Enhanced rate-distortion optimization for stereo interleaving video coding","authors":"Qian Ma, Yongbing Zhang, Qiong Liu, Xiangyang Ji, Qionghai Dai","doi":"10.1109/IC3D.2011.6584364","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584364","url":null,"abstract":"Stereo interleaving format attracts more and more attention due to its back compatibility with all existing 2D video coding standards as well as its high efficiency of stereoscopic video compression. As one of the most significant coding components, rate-distortion optimization (RDO) in stereo interleaving video coding does not take into account its application scenarios where the reconstructed video needs up-sampling for displaying. To promote the efficiency of stereo interleaving video compression, we propose an enhanced RDO where the up-sampling is taken into consideration on distortion measurement. The proposed algorithm changes nothing concerning the syntax and is compatible with current decoders. Experimental results demonstrate that the proposed enhanced RDO is able to reduce the bitrates by 10%-44% on average, and achieve 0.1-0.65dB gain on PSNR compared with conventional RDO.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125100305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584392
F. Klose, C. Lipski, K. Ruhl, B. Meyer, M. Magnor
We present our toolchain of free-viewpoint video and dynamic scene reconstruction for video and stereoscopic content creation. Our tools take video data from set of sparse unsynchronized cameras and give great freedom during the post production. From input data we can either generate new viewpoints employing the Virtual Video Camera a purely image based system or generate 3D scene models. The approaches are explained and guidelines for weighing their specific advantages and disadvantages in respect to a concrete application are given.
{"title":"A toolchain for capturing and rendering stereo and multi-view datasets","authors":"F. Klose, C. Lipski, K. Ruhl, B. Meyer, M. Magnor","doi":"10.1109/IC3D.2011.6584392","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584392","url":null,"abstract":"We present our toolchain of free-viewpoint video and dynamic scene reconstruction for video and stereoscopic content creation. Our tools take video data from set of sparse unsynchronized cameras and give great freedom during the post production. From input data we can either generate new viewpoints employing the Virtual Video Camera a purely image based system or generate 3D scene models. The approaches are explained and guidelines for weighing their specific advantages and disadvantages in respect to a concrete application are given.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127681374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}