Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032580
P. Aflaki, Maryam Homayouni, M. Hannuksela, M. Gabbouj
In this paper, a coding scheme targeting stereoscopic content for polarized displays is introduced. It is proposed to use row-interleaved sampling of the views. Asymmetry is achieved by selection of odd/even rows for different views based on the format they will be shown on a polarized display. Coding performance of several different multiview coding schemes with inter-view prediction was analyzed and compared with the anchor case where there is no downsampling applied to the input content. The objective results show that the proposed row-interleaved sampling scheme outperforms all other schemes.
{"title":"Row-interleaved sampling for stereoscopic video coding targeting polarized displays","authors":"P. Aflaki, Maryam Homayouni, M. Hannuksela, M. Gabbouj","doi":"10.1109/IC3D.2014.7032580","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032580","url":null,"abstract":"In this paper, a coding scheme targeting stereoscopic content for polarized displays is introduced. It is proposed to use row-interleaved sampling of the views. Asymmetry is achieved by selection of odd/even rows for different views based on the format they will be shown on a polarized display. Coding performance of several different multiview coding schemes with inter-view prediction was analyzed and compared with the anchor case where there is no downsampling applied to the input content. The objective results show that the proposed row-interleaved sampling scheme outperforms all other schemes.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125663297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032585
M. Hasan, J. Arnold, M. Frater
Broadcasting of high definition stereoscopic 3D videos is growing rapidly because of greater demand in the mass consumer market. In spite of increasing consumer interest, poor quality, crosstalk or side effects and visual quality degradation due to packet loss during transmission has hampered the advancement of 3D visualization. The quality assessment of distorted 3D video is a crucial element in designing and arranging advanced immersive media distribution platforms. A widely accepted no-reference quality metric to measure 3D video considering the human visual system (HVS) is yet to be developed. In this paper we have proposed a quality assessment (QA) criterion that can be measured without the original video. At first, we proposed a disparity index, that is measured by region based similarity matching and then edge magnitude difference is detected for visually significant areas of the image. Finally, an assessment metric is generated to measure the 3D videos focusing on human perception. Experimental analysis with common video datasets and comparison with different algorithms shows the efficiency of the proposed algorithm for 3D stereoscopic videos in terms of perceptual characteristics.
{"title":"No-reference quality assessment of 3D videos based on human visual perception","authors":"M. Hasan, J. Arnold, M. Frater","doi":"10.1109/IC3D.2014.7032585","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032585","url":null,"abstract":"Broadcasting of high definition stereoscopic 3D videos is growing rapidly because of greater demand in the mass consumer market. In spite of increasing consumer interest, poor quality, crosstalk or side effects and visual quality degradation due to packet loss during transmission has hampered the advancement of 3D visualization. The quality assessment of distorted 3D video is a crucial element in designing and arranging advanced immersive media distribution platforms. A widely accepted no-reference quality metric to measure 3D video considering the human visual system (HVS) is yet to be developed. In this paper we have proposed a quality assessment (QA) criterion that can be measured without the original video. At first, we proposed a disparity index, that is measured by region based similarity matching and then edge magnitude difference is detected for visually significant areas of the image. Finally, an assessment metric is generated to measure the 3D videos focusing on human perception. Experimental analysis with common video datasets and comparison with different algorithms shows the efficiency of the proposed algorithm for 3D stereoscopic videos in terms of perceptual characteristics.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126605915","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032576
Tomás Majtner, D. Svoboda
The recognition of patterns with focus on texture and shape analysis is still very hot topic, especially in biomédical image processing. In this article, we introduce 3D extensions of well-known approaches for this particular area. We focus on the collection of MPEG-7 image descriptors, specifically on the Edge Histogram Descriptor (EHD) and Gabor features, which are the core of the Homogeneous Texture Descriptor (HTD). The proposed extensions are evaluated on the dataset consisting of three classes of 3D volumetric biomédical images. Two different classifiers, namely k-NN and Multi-Class SVM, are used to evaluate the proposed algorithms. According to the presented tests, the proposed 3D extensions clearly outperform their 2D equivalents in the classification tasks.
{"title":"Texture analysis using 3D Gabor features and 3D MPEG-7 Edge Histogram descriptor in fluorescence microscopy","authors":"Tomás Majtner, D. Svoboda","doi":"10.1109/IC3D.2014.7032576","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032576","url":null,"abstract":"The recognition of patterns with focus on texture and shape analysis is still very hot topic, especially in biomédical image processing. In this article, we introduce 3D extensions of well-known approaches for this particular area. We focus on the collection of MPEG-7 image descriptors, specifically on the Edge Histogram Descriptor (EHD) and Gabor features, which are the core of the Homogeneous Texture Descriptor (HTD). The proposed extensions are evaluated on the dataset consisting of three classes of 3D volumetric biomédical images. Two different classifiers, namely k-NN and Multi-Class SVM, are used to evaluate the proposed algorithms. According to the presented tests, the proposed 3D extensions clearly outperform their 2D equivalents in the classification tasks.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117013210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032572
Mohan Liu, K. Müller
This paper presents an efficient approach to measure sharpness mismatches between stereoscopic views. Sharpness mismatch can occur through focus mismatches between stereoscopic cameras, errors in post-processing or even for low-bandwidth transmission, where one view is subsampled or transmitted at a much lower rate. This artifact can lead to a degraded 3D experience for observers. In this paper, the sharpness mismatch score is estimated by measuring the width deviations of edge pairs in each valid depth plane. The mismatch probability is then calculated considering the perceptibility of edge width deviations. In the experiments, Gaussian low-pass filters were used to generate global sharpness mismatches between stereoscopic views since defocus-based effects of lens aberrations can be modeled as Gaussian blur. Thus, the global sharpness distortions simulate the focus mismatch of stereo cameras. The disparity maps of test videos were automatically generated and corrected. In addition, original high-quality disparity maps of the test datasets were used as benchmarks. According to the experimental results, we show that the proposed approach performs well on measuring sharpness mismatch between stereoscopic views by comparison with some state-of-the-art metrics.
{"title":"Automatic analysis of sharpness mismatch between stereoscopic views for stereo 3D videos","authors":"Mohan Liu, K. Müller","doi":"10.1109/IC3D.2014.7032572","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032572","url":null,"abstract":"This paper presents an efficient approach to measure sharpness mismatches between stereoscopic views. Sharpness mismatch can occur through focus mismatches between stereoscopic cameras, errors in post-processing or even for low-bandwidth transmission, where one view is subsampled or transmitted at a much lower rate. This artifact can lead to a degraded 3D experience for observers. In this paper, the sharpness mismatch score is estimated by measuring the width deviations of edge pairs in each valid depth plane. The mismatch probability is then calculated considering the perceptibility of edge width deviations. In the experiments, Gaussian low-pass filters were used to generate global sharpness mismatches between stereoscopic views since defocus-based effects of lens aberrations can be modeled as Gaussian blur. Thus, the global sharpness distortions simulate the focus mismatch of stereo cameras. The disparity maps of test videos were automatically generated and corrected. In addition, original high-quality disparity maps of the test datasets were used as benchmarks. According to the experimental results, we show that the proposed approach performs well on measuring sharpness mismatch between stereoscopic views by comparison with some state-of-the-art metrics.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"168 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123276501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032596
A. Lejeune, David Grogna, Marc Van Droogenbroeck, J. Verly
Many applications require the use of multiple cameras to cover a large volume. In this paper, we evaluate several pairwise calibration techniques dedicated to multiple range cameras. We compare the precision of a self-calibration technique based on the movement in front of the cameras to object based calibration. While the self-calibration technique is less precise than its counterparts, it yields a first estimation of the transformation between the cameras and permits to detect when the cameras become mis-aligned. Therefore, this technique is useful in a practical situations.
{"title":"Evaluation of pairwise calibration techniques for range cameras and their ability to detect a misalignment","authors":"A. Lejeune, David Grogna, Marc Van Droogenbroeck, J. Verly","doi":"10.1109/IC3D.2014.7032596","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032596","url":null,"abstract":"Many applications require the use of multiple cameras to cover a large volume. In this paper, we evaluate several pairwise calibration techniques dedicated to multiple range cameras. We compare the precision of a self-calibration technique based on the movement in front of the cameras to object based calibration. While the self-calibration technique is less precise than its counterparts, it yields a first estimation of the transformation between the cameras and permits to detect when the cameras become mis-aligned. Therefore, this technique is useful in a practical situations.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134100595","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032579
A. Dorado, G. Saavedra, Seokmin Hong, M. Martínez-Corral
We report a new algorithm for the generation of the microimages ready for their projection into an integral imaging monitor. The algorithm is based in the transformation properties of the plenoptic field captured with an array of digital cameras. We show that a small number of cameras can produce the microimages for displaying 3D scenes with resolution and parallax fully adapted to the monitor features.
{"title":"Computation of microimages for plenoptic display","authors":"A. Dorado, G. Saavedra, Seokmin Hong, M. Martínez-Corral","doi":"10.1109/IC3D.2014.7032579","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032579","url":null,"abstract":"We report a new algorithm for the generation of the microimages ready for their projection into an integral imaging monitor. The algorithm is based in the transformation properties of the plenoptic field captured with an array of digital cameras. We show that a small number of cameras can produce the microimages for displaying 3D scenes with resolution and parallax fully adapted to the monitor features.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121905112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032591
Marc Braham, A. Lejeune, Marc Van Droogenbroeck
This paper proposes a new pixel-based background subtraction technique, applicable to range images, to detect motion. Our method exploits the physical meaning of depth information, which leads to an improved background/foreground segmentation and the instantaneous suppression of ghosts that would appear on color images. In particular, our technique considers certain characteristics of depth measurements, such as failures for certain pixels or the non-uniformity of the spatial distribution of noise in range images, to build an improved pixel-based background model. Experiments show that incorporating specificities related to depth measurements allows us to propose a method whose performance is increased with respect to other state-of-the-art methods.
{"title":"A physically motivated pixel-based model for background subtraction in 3D images","authors":"Marc Braham, A. Lejeune, Marc Van Droogenbroeck","doi":"10.1109/IC3D.2014.7032591","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032591","url":null,"abstract":"This paper proposes a new pixel-based background subtraction technique, applicable to range images, to detect motion. Our method exploits the physical meaning of depth information, which leads to an improved background/foreground segmentation and the instantaneous suppression of ghosts that would appear on color images. In particular, our technique considers certain characteristics of depth measurements, such as failures for certain pixels or the non-uniformity of the spatial distribution of noise in range images, to build an improved pixel-based background model. Experiments show that incorporating specificities related to depth measurements allows us to propose a method whose performance is increased with respect to other state-of-the-art methods.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130196416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032578
Min-Hung Chen, Ching-Fan Chiang, Yi-Chang Lu
Depth estimation is one of the new functions provided by hand-held light field cameras. However, the quality of depth estimation is very sensitive to noise, which is especially a problem for scenes under low light conditions. In this paper, we propose a depth estimation flow for light field data, which can be fully-automated and no noise characteristics are required a priori. The results of Root Mean Square Error (RMSE) and Percentage of Bad Matching Pixels (PBM) show the effectiveness of this iterative correlation-based depth estimation flow even with basic filtering functions.
{"title":"Depth estimation for hand-held light field cameras under low light conditions","authors":"Min-Hung Chen, Ching-Fan Chiang, Yi-Chang Lu","doi":"10.1109/IC3D.2014.7032578","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032578","url":null,"abstract":"Depth estimation is one of the new functions provided by hand-held light field cameras. However, the quality of depth estimation is very sensitive to noise, which is especially a problem for scenes under low light conditions. In this paper, we propose a depth estimation flow for light field data, which can be fully-automated and no noise characteristics are required a priori. The results of Root Mean Square Error (RMSE) and Percentage of Bad Matching Pixels (PBM) show the effectiveness of this iterative correlation-based depth estimation flow even with basic filtering functions.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122957119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032588
Yao Yao, Hao Zhu, Yongming Nie, X. Ji, Xun Cao
Optical flow estimation is one of the popular methods to obtain the depth maps in multi-view stereo due to its high accuracy and robustness. In traditional optical flow estimation, the energy function contains three assumptions: intensity constancy assumption, gradient constancy assumption, and global smoothness assumption. In this work, we propose a local smoothness assumption to constrain the optical flow disparity in neighboring pixels. We first study the new smoothness term and its corresponding energy function, and present a practical iteration approach to minimize the energy function. Later we apply this new estimation method to the multi-view stereo system and obtain the depth maps of different image pairs. Our results demonstrate the good performance of the algorithm in acquiring smoothing surface when comparing to the traditional methods.
{"title":"Revised depth map estimation for multi-view stereo","authors":"Yao Yao, Hao Zhu, Yongming Nie, X. Ji, Xun Cao","doi":"10.1109/IC3D.2014.7032588","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032588","url":null,"abstract":"Optical flow estimation is one of the popular methods to obtain the depth maps in multi-view stereo due to its high accuracy and robustness. In traditional optical flow estimation, the energy function contains three assumptions: intensity constancy assumption, gradient constancy assumption, and global smoothness assumption. In this work, we propose a local smoothness assumption to constrain the optical flow disparity in neighboring pixels. We first study the new smoothness term and its corresponding energy function, and present a practical iteration approach to minimize the energy function. Later we apply this new estimation method to the multi-view stereo system and obtain the depth maps of different image pairs. Our results demonstrate the good performance of the algorithm in acquiring smoothing surface when comparing to the traditional methods.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115157155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/IC3D.2014.7032589
Chun-Chang Yu, Chia-Hao Cheng, Pei-Chun Lin, C. C. Chen
This paper focuses on the visual fatigue issue while viewing 3D contents. The issue is caused by the distance between the screen and the fused images. A stereo image depth optimization system with disparity map calculation, viewpoint optimization and stereo image synthesis is proposed to solve the issue with the following procedure: first, its disparity map calculation adopts the modified binary window block matching algorithm so that the complex and iterative computations can be accelerated by hardware implementation strategies including parallel color difference calculation, parallel memory banks, window shift, and pipelined architecture; second, the viewpoint optimization modifies disparities to the zone of comfort; third, stereo images are synthesized through Depth-Image-Based-Rendering (DIBR); finally, the stereo image depth optimization system is realized on the FPGA board and video files are shown via the HDMI interface. This hardware implementation turns out to be more cost-efficient to achieve high-speed performance when compared with previous works.
本文主要研究观看3D内容时的视觉疲劳问题。这个问题是由屏幕和融合图像之间的距离引起的。针对立体图像深度优化问题,提出了一种视差图计算、视点优化和立体图像合成的立体图像深度优化系统。首先,视差图计算采用改进的二进制窗口块匹配算法,通过并行色差计算、并行存储库、窗口移位和流水线架构等硬件实现策略加快了复杂的迭代计算;第二,视点优化将视差调整到舒适区;第三,通过深度图像渲染(deep - image - based rendering, DIBR)合成立体图像;最后,在FPGA板上实现了立体图像深度优化系统,并通过HDMI接口显示视频文件。与以前的工作相比,这种硬件实现在实现高速性能方面更具成本效益。
{"title":"Cost-efficient hardware implementation of stereo image depth optimization system","authors":"Chun-Chang Yu, Chia-Hao Cheng, Pei-Chun Lin, C. C. Chen","doi":"10.1109/IC3D.2014.7032589","DOIUrl":"https://doi.org/10.1109/IC3D.2014.7032589","url":null,"abstract":"This paper focuses on the visual fatigue issue while viewing 3D contents. The issue is caused by the distance between the screen and the fused images. A stereo image depth optimization system with disparity map calculation, viewpoint optimization and stereo image synthesis is proposed to solve the issue with the following procedure: first, its disparity map calculation adopts the modified binary window block matching algorithm so that the complex and iterative computations can be accelerated by hardware implementation strategies including parallel color difference calculation, parallel memory banks, window shift, and pipelined architecture; second, the viewpoint optimization modifies disparities to the zone of comfort; third, stereo images are synthesized through Depth-Image-Based-Rendering (DIBR); finally, the stereo image depth optimization system is realized on the FPGA board and video files are shown via the HDMI interface. This hardware implementation turns out to be more cost-efficient to achieve high-speed performance when compared with previous works.","PeriodicalId":244221,"journal":{"name":"2014 International Conference on 3D Imaging (IC3D)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122631510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}