Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584381
P. Aflaki, D. Rusanovskyy, T. Utriainen, E. Pesonen, M. Hannuksela, Satu Jumisko-Pyykkö, M. Gabbouj
Depth-enhanced multiview video formats, such as the multiview video plus depth (MVD) format, enable a natural 3D visual experience which cannot be brought by traditional 2D or stereo video services. In this paper we studied an asymmetric MVD technique for coding of three views that enabled rendering of the same bitstream on stereoscopic displays and multiview autostereoscopic displays. A larger share of bitrate was allocated to a central view, whereas two side views were coded at lower quality. The three decoded views were used by a Depth-Image-Based Rendering algorithm (DIBR) to produce virtual intermediate views. A stereopair at a suitable separation for viewing on a stereoscopic display was selected among the synthesized views. A large-scale subjective assessment of the selected synthesized stereopair was performed. A bitrate reduction of 20% on average and up to 22% was achieved with no penalties on subjective perceived quality. In addition, our analysis shows that a similar bitrate reduction gain with no difference in subjective quality can be achieved in multiview autostereoscopic display scenario.
深度增强的多视图视频格式,如多视图视频加深度(MVD)格式,可以实现传统2D或立体视频服务无法带来的自然3D视觉体验。在本文中,我们研究了一种用于三视图编码的非对称MVD技术,该技术能够在立体显示器和多视图自立体显示器上呈现相同的比特流。更大的比特率分配给中心视图,而两个侧视图编码的质量较低。通过深度图像渲染算法(deep - image - based Rendering algorithm, DIBR)生成虚拟中间视图。在合成视图中选择适合在立体显示器上观看的距离的立体对。对所选择的合成立体对进行了大规模的主观评价。比特率平均降低了20%,最高可降低22%,而主观感知质量没有受到任何影响。此外,我们的分析表明,在多视角自动立体显示场景中,可以实现类似的比特率降低增益,而主观质量没有差异。
{"title":"Study of asymmetric quality between coded views in depth-enhanced multiview video coding","authors":"P. Aflaki, D. Rusanovskyy, T. Utriainen, E. Pesonen, M. Hannuksela, Satu Jumisko-Pyykkö, M. Gabbouj","doi":"10.1109/IC3D.2011.6584381","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584381","url":null,"abstract":"Depth-enhanced multiview video formats, such as the multiview video plus depth (MVD) format, enable a natural 3D visual experience which cannot be brought by traditional 2D or stereo video services. In this paper we studied an asymmetric MVD technique for coding of three views that enabled rendering of the same bitstream on stereoscopic displays and multiview autostereoscopic displays. A larger share of bitrate was allocated to a central view, whereas two side views were coded at lower quality. The three decoded views were used by a Depth-Image-Based Rendering algorithm (DIBR) to produce virtual intermediate views. A stereopair at a suitable separation for viewing on a stereoscopic display was selected among the synthesized views. A large-scale subjective assessment of the selected synthesized stereopair was performed. A bitrate reduction of 20% on average and up to 22% was achieved with no penalties on subjective perceived quality. In addition, our analysis shows that a similar bitrate reduction gain with no difference in subjective quality can be achieved in multiview autostereoscopic display scenario.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126042381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584368
B. Penelle, Arnaud Schenkel, N. Warzée
A RGB-D image combines, for each pixel, the classical three color channels with a fourth channel providing depth information. Devices that produce RGB-D images in real time with a rather good resolution are currently available on the market. With this type of device, it is possible to acquire and to process, in real time, 3D textured information, paving the way for numerous applications in the field of computer imaging and vision. In this paper, we analyse the accuracy of a low cost system and we see how this kind of device and the RGB-D images it produces allow us to acquire 3D models of real objects. A first application is presented that combines multiple RGB-D images of a static scene, taken from different viewpoints, in order to reconstruct a complete 3D model of the scene. A second application combines on-the-fly RGB-D images coming from multiple devices, generating a 3D model where the problems of occlusions inherent in monocular observations are drastically reduced.
{"title":"Geometrical 3D reconstruction using real-time RGB-D cameras","authors":"B. Penelle, Arnaud Schenkel, N. Warzée","doi":"10.1109/IC3D.2011.6584368","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584368","url":null,"abstract":"A RGB-D image combines, for each pixel, the classical three color channels with a fourth channel providing depth information. Devices that produce RGB-D images in real time with a rather good resolution are currently available on the market. With this type of device, it is possible to acquire and to process, in real time, 3D textured information, paving the way for numerous applications in the field of computer imaging and vision. In this paper, we analyse the accuracy of a low cost system and we see how this kind of device and the RGB-D images it produces allow us to acquire 3D models of real objects. A first application is presented that combines multiple RGB-D images of a static scene, taken from different viewpoints, in order to reconstruct a complete 3D model of the scene. A second application combines on-the-fly RGB-D images coming from multiple devices, generating a 3D model where the problems of occlusions inherent in monocular observations are drastically reduced.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"362 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126700065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584369
P. Surman, S. Day, E. Willman, H. Baghsiahi, I. Sexton, K. Hopf
This paper describes the development of an autostereoscopic laser-based display that can serve several users who are able to move freely over a large area; it is intended for television applications and operates by tracking the positions of the viewers' heads and directing regions referred to as exit pupils towards their eyes. An exit pupil is a region where either a left or a right image is seen across the complete area of the screen. A description of the 60 and 120Hz prototypes developed is given. A set-up that does not show images but demonstrates the operation of the novel dynamic exit pupil formation system is also described,
{"title":"HELIUM3D: A laser-scanned head-tracked autostereoscopic display","authors":"P. Surman, S. Day, E. Willman, H. Baghsiahi, I. Sexton, K. Hopf","doi":"10.1109/IC3D.2011.6584369","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584369","url":null,"abstract":"This paper describes the development of an autostereoscopic laser-based display that can serve several users who are able to move freely over a large area; it is intended for television applications and operates by tracking the positions of the viewers' heads and directing regions referred to as exit pupils towards their eyes. An exit pupil is a region where either a left or a right image is seen across the complete area of the screen. A description of the 60 and 120Hz prototypes developed is given. A set-up that does not show images but demonstrates the operation of the novel dynamic exit pupil formation system is also described,","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125251487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584365
Hao Cheng, P. An, Hejian Li, Zhaoyang Zhang
This paper introduces a stereo image rectification algorithm based on image feature points for multi-view three-dimension (3D) display system. First, we mark the feature points in multi-view images, and calculate the rectification parameters. Then we use these parameters for rotating image and shifting image in vertical direction in order to eliminate the vertical parallax of multi-view images. Finally according to zero-parallax setting (ZPS), we adjust the horizontal parallax to get a better stereo image for multi-view 3D display. After the multi-view information is corrected, the stereo effect has been greatly enhanced in multi-view display system. The algorithm has low complexity in real-time 3D system and can improve the stereo image from the point of the observer.
{"title":"Stereo image rectification algorithm for multi-view 3D display","authors":"Hao Cheng, P. An, Hejian Li, Zhaoyang Zhang","doi":"10.1109/IC3D.2011.6584365","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584365","url":null,"abstract":"This paper introduces a stereo image rectification algorithm based on image feature points for multi-view three-dimension (3D) display system. First, we mark the feature points in multi-view images, and calculate the rectification parameters. Then we use these parameters for rotating image and shifting image in vertical direction in order to eliminate the vertical parallax of multi-view images. Finally according to zero-parallax setting (ZPS), we adjust the horizontal parallax to get a better stereo image for multi-view 3D display. After the multi-view information is corrected, the stereo effect has been greatly enhanced in multi-view display system. The algorithm has low complexity in real-time 3D system and can improve the stereo image from the point of the observer.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132278322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584384
S. Mattoccia, Leonardo De-Maeztu
Recent cost aggregation strategies that adapt their weights to image content enabled local algorithms to obtain results comparable to those of global algorithms based on more complex disparity optimization methods. Unfortunately, despite the potential advantages in terms of memory footprint and algorithmic simplicity compared to global algorithms, most of the state-of-the-art cost aggregation strategies deployed in local algorithms are extremely slow. In fact, their execution time is comparable and often worse than those of global approaches. In this paper we propose a framework for accurate and fast cost aggregation based on segmentation that allows us to obtain results comparable to state-of-the-art approaches much more efficiently (the execution time drops from minutes to seconds). A further speed-up is achieved taking advantage of multi-core capabilities available nowadays in almost any processor. The comparison with state-of-the-art cost aggregation strategies highlights the effectiveness of our proposal.
{"title":"A fast segmentation-driven algorithm for accurate stereo correspondence","authors":"S. Mattoccia, Leonardo De-Maeztu","doi":"10.1109/IC3D.2011.6584384","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584384","url":null,"abstract":"Recent cost aggregation strategies that adapt their weights to image content enabled local algorithms to obtain results comparable to those of global algorithms based on more complex disparity optimization methods. Unfortunately, despite the potential advantages in terms of memory footprint and algorithmic simplicity compared to global algorithms, most of the state-of-the-art cost aggregation strategies deployed in local algorithms are extremely slow. In fact, their execution time is comparable and often worse than those of global approaches. In this paper we propose a framework for accurate and fast cost aggregation based on segmentation that allows us to obtain results comparable to state-of-the-art approaches much more efficiently (the execution time drops from minutes to seconds). A further speed-up is achieved taking advantage of multi-core capabilities available nowadays in almost any processor. The comparison with state-of-the-art cost aggregation strategies highlights the effectiveness of our proposal.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"1993 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125539797","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584370
Guannan Li, Yebin Liu, Qionghai Dai
We present an approach to reconstruct 3D fine-scale surface models for non-Lambertian objects from multi-view multi-illumination image sets. Unlike most previous work in photometric stereo, this approach works for general lighting conditions, i.e. natural outdoor illumination. Our method begins with a raw 3D model reconstructed from available multi-view stereo techniques. Considering the sparse characteristics of surface reflectance in the view-illumination space, we first estimate the diffuse appearance of the 3D model from the multiview captured images, and then refine it using the surface appearance under varying illuminations. With the separated low rank diffuse component, we exploit the photometric cues to recover detailed surface structure. Experimental results on various real world scenes validate that the proposed method is able to handle surfaces with specular reflectance even including saturated colours, highlight and cast-shadows.
{"title":"Multi-view photometric stereo of non-Lambertian surface under general illuminations","authors":"Guannan Li, Yebin Liu, Qionghai Dai","doi":"10.1109/IC3D.2011.6584370","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584370","url":null,"abstract":"We present an approach to reconstruct 3D fine-scale surface models for non-Lambertian objects from multi-view multi-illumination image sets. Unlike most previous work in photometric stereo, this approach works for general lighting conditions, i.e. natural outdoor illumination. Our method begins with a raw 3D model reconstructed from available multi-view stereo techniques. Considering the sparse characteristics of surface reflectance in the view-illumination space, we first estimate the diffuse appearance of the 3D model from the multiview captured images, and then refine it using the surface appearance under varying illuminations. With the separated low rank diffuse component, we exploit the photometric cues to recover detailed surface structure. Experimental results on various real world scenes validate that the proposed method is able to handle surfaces with specular reflectance even including saturated colours, highlight and cast-shadows.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131494951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584363
O. Cahen
Floating window means projecting the window ahead from the screen. Benefits, methods, requirements for comfortable viewing conditions, are reviewed.
浮动窗口是指将窗口从屏幕上伸出来。对舒适观看条件的好处、方法和要求进行了综述。
{"title":"The floating window, its benefits, methods, requirements","authors":"O. Cahen","doi":"10.1109/IC3D.2011.6584363","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584363","url":null,"abstract":"Floating window means projecting the window ahead from the screen. Benefits, methods, requirements for comfortable viewing conditions, are reviewed.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132420698","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584383
Christophe de Dinechin, Catherine Burvelle, Jerome Forissier
Tao Presentations uses a dialect of the XL programming language to describe interactive multimedia 3D documents. This approach makes it easy to create 3D contents that can be used to present information, to visualize scientific data or to explore stereoscopic or auto-stereoscopic effects rapidly. The demands of an interactive, real-time environment have created a number of interesting challenges for us to solve, ranging from language expressiveness and document semantics to graphics performance and rendering quality.
Tao Presentations使用XL编程语言的一种方言来描述交互式多媒体3D文档。这种方法可以很容易地创建3D内容,用于呈现信息、可视化科学数据或快速探索立体或自动立体效果。交互式、实时环境的需求给我们带来了许多有趣的挑战,从语言表现力和文档语义到图形性能和渲染质量。
{"title":"Tao, a 3D dynamic document description language","authors":"Christophe de Dinechin, Catherine Burvelle, Jerome Forissier","doi":"10.1109/IC3D.2011.6584383","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584383","url":null,"abstract":"Tao Presentations uses a dialect of the XL programming language to describe interactive multimedia 3D documents. This approach makes it easy to create 3D contents that can be used to present information, to visualize scientific data or to explore stereoscopic or auto-stereoscopic effects rapidly. The demands of an interactive, real-time environment have created a number of interesting challenges for us to solve, ranging from language expressiveness and document semantics to graphics performance and rendering quality.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133540433","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584387
P. K. Kumar, T. Kumar
Through the wall sensing is seeing a rapid growth with the present day technology. Particularly Ultra Wideband (UWB) technology paves the way for this. Three dimensional imaging of humans behind the walls, foliage or any other rubble gives information which can help to save lives. This paper simulates a human like structure with the environment of being behind wall and extracts the three dimensional imaging of the human structure. Electromagnetic signals were transmitted and based on the received echoes after processing a 3D Imaging is obtained. Signal processing aspects before imaging and the method used for obtaining the tree dimensional imaging is also discussed.
{"title":"Three dimensional imaging for through-the-wall human sensing","authors":"P. K. Kumar, T. Kumar","doi":"10.1109/IC3D.2011.6584387","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584387","url":null,"abstract":"Through the wall sensing is seeing a rapid growth with the present day technology. Particularly Ultra Wideband (UWB) technology paves the way for this. Three dimensional imaging of humans behind the walls, foliage or any other rubble gives information which can help to save lives. This paper simulates a human like structure with the environment of being behind wall and extracts the three dimensional imaging of the human structure. Electromagnetic signals were transmitted and based on the received echoes after processing a 3D Imaging is obtained. Signal processing aspects before imaging and the method used for obtaining the tree dimensional imaging is also discussed.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131158869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/IC3D.2011.6584378
Mårten Sjöström, R. Olsson, Rolf Dalin
The use of digital signs simplifies distribution. Importantly, it draws more attention than static signs. A way to increase attention is to add an experienced depth. The paper discusses possible alternatives for extending an existing digital signage system to display stereoscopic 3D contents, comparing a decentralized distribution solution and a centralized solution. A functional prototype system was implemented. A new 3D player was developed to render views from different formats. The implemented system was used to study customer behavior when exposed to digital stereoscopic 3D signage in a direct sales situation. The proportion of sales of selected products related to the total number of sold products varied approximately equally before and during tests. An interview study suggests that the sign did not interact with customer decisions: customers were lost at different stages in this series of steps, among others the sign placement.
{"title":"A digital 3D signage system and its effect on customer behavior","authors":"Mårten Sjöström, R. Olsson, Rolf Dalin","doi":"10.1109/IC3D.2011.6584378","DOIUrl":"https://doi.org/10.1109/IC3D.2011.6584378","url":null,"abstract":"The use of digital signs simplifies distribution. Importantly, it draws more attention than static signs. A way to increase attention is to add an experienced depth. The paper discusses possible alternatives for extending an existing digital signage system to display stereoscopic 3D contents, comparing a decentralized distribution solution and a centralized solution. A functional prototype system was implemented. A new 3D player was developed to render views from different formats. The implemented system was used to study customer behavior when exposed to digital stereoscopic 3D signage in a direct sales situation. The proportion of sales of selected products related to the total number of sold products varied approximately equally before and during tests. An interview study suggests that the sign did not interact with customer decisions: customers were lost at different stages in this series of steps, among others the sign placement.","PeriodicalId":395174,"journal":{"name":"2011 International Conference on 3D Imaging (IC3D)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124344258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}