首页 > 最新文献

IVMSP 2013最新文献

英文 中文
Study of a FPGA real-time multi-cameras cylindrical panorama video system with low latency and high performance FPGA低延迟高性能实时多摄像头圆柱全景视频系统的研究
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611890
Yuan Xu, Qinghai Zhou, Liwei Gong, Mingcheng Zhu, Robert K. F. Teng
This paper presents a method of real-time panoramic video system implemented by using FPGA. In order to meet the real-time and high-resolution requirements, a novel method based on pipeline CORDIC and look-up tables for simultaneous image transformations, such as lens distortion correction, perspective transformation and image cylinder projection has been developed. Taking the advantages of the high speed parallel architecture of FPGA and the developed novel algorithm modules, a real-time, low latency and high-performance panoramic video has been studied and developed. The video system presented in this paper could generate panoramic video of 5*1280(H)*720(V) @15FPS or 5*1280(H)*360(V) @30FPS in real-time. In order to display the panoramic video, a display module that can output the panoramic video in slide mode has also been built.
提出了一种利用FPGA实现实时全景视频系统的方法。为了满足实时性和高分辨率的要求,提出了一种基于流水线CORDIC和查找表的图像同步变换方法,实现了镜头畸变校正、透视变换和图像柱面投影等图像同步变换。利用FPGA的高速并行架构和所开发的新型算法模块,研究开发了实时、低延迟、高性能的全景视频。本文提出的视频系统可以实时生成5*1280(H)*720(V) @15FPS或5*1280(H)*360(V) @30FPS的全景视频。为了显示全景视频,还构建了一个可以以滑动方式输出全景视频的显示模块。
{"title":"Study of a FPGA real-time multi-cameras cylindrical panorama video system with low latency and high performance","authors":"Yuan Xu, Qinghai Zhou, Liwei Gong, Mingcheng Zhu, Robert K. F. Teng","doi":"10.1109/IVMSPW.2013.6611890","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611890","url":null,"abstract":"This paper presents a method of real-time panoramic video system implemented by using FPGA. In order to meet the real-time and high-resolution requirements, a novel method based on pipeline CORDIC and look-up tables for simultaneous image transformations, such as lens distortion correction, perspective transformation and image cylinder projection has been developed. Taking the advantages of the high speed parallel architecture of FPGA and the developed novel algorithm modules, a real-time, low latency and high-performance panoramic video has been studied and developed. The video system presented in this paper could generate panoramic video of 5*1280(H)*720(V) @15FPS or 5*1280(H)*360(V) @30FPS in real-time. In order to display the panoramic video, a display module that can output the panoramic video in slide mode has also been built.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127041987","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Perceptual preference of S3D over 2D for HDTV in dependence of video quality and depth 高清电视中S3D对2D的感知偏好与视频质量和深度的关系
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611905
Pierre R. Lebreton, A. Raake, M. Barkowsky, P. Callet
3D video quality of experience (QoE) is a multidimensional problem and many factors contribute to the global experience by the user. Due to this multidimensionality, this paper evaluates the integral 3D video QoE and relates it with image quality and depth. Subjective tests have been conducted using paired comparison to evaluate 3D QoE and the preference of 3D over 2D with different combinations of coding conditions. Depth scores were available from previous work and were used to check their relation with 3DQoE; the difference between 2D and 3D QoE is found to be a function of the picture quality, and the desired preference of 3D presentation over 2D can be reached when pictorial quality is high enough (VQM score lower than 0.24). A factor ranging from 0.08 to 0.76 with a mean of 0.71 between pictorial quality and preference of 3D was also found.
3D视频体验质量(QoE)是一个多维度的问题,影响用户整体体验的因素很多。由于这种多维性,本文对三维视频的整体质量质量进行了评价,并将其与图像质量和深度联系起来。使用配对比较进行了主观测试,以评估3D QoE以及在不同编码条件组合下3D优于2D的偏好。深度分数从以前的工作中获得,并用于检查其与3DQoE的关系;发现2D和3D QoE的差异是图像质量的函数,当图像质量足够高时(VQM得分低于0.24),3D呈现优于2D呈现。还发现图像质量与3D偏好之间的因子范围为0.08至0.76,平均值为0.71。
{"title":"Perceptual preference of S3D over 2D for HDTV in dependence of video quality and depth","authors":"Pierre R. Lebreton, A. Raake, M. Barkowsky, P. Callet","doi":"10.1109/IVMSPW.2013.6611905","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611905","url":null,"abstract":"3D video quality of experience (QoE) is a multidimensional problem and many factors contribute to the global experience by the user. Due to this multidimensionality, this paper evaluates the integral 3D video QoE and relates it with image quality and depth. Subjective tests have been conducted using paired comparison to evaluate 3D QoE and the preference of 3D over 2D with different combinations of coding conditions. Depth scores were available from previous work and were used to check their relation with 3DQoE; the difference between 2D and 3D QoE is found to be a function of the picture quality, and the desired preference of 3D presentation over 2D can be reached when pictorial quality is high enough (VQM score lower than 0.24). A factor ranging from 0.08 to 0.76 with a mean of 0.71 between pictorial quality and preference of 3D was also found.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125750897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
View-independent human action recognition based on multi-view action images and discriminant learning 基于多视点动作图像和判别学习的非视点人类动作识别
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611931
Alexandros Iosifidis, A. Tefas, I. Pitas
In this paper a novel view-independent human action recognition method is proposed. A multi-camera setup is used to capture the human body from different viewing angles. Actions are described by a novel action representation, the so-called multi-view action image (MVAI), which effectively addresses the camera viewpoint identification problem, i.e., the identification of the position of each camera with respect to the person's body. Linear Discriminant Analysis is applied on the MVAIs in order to to map actions to a discriminant feature space where actions are classified by using a simple nearest class centroid classification scheme. Experimental results denote the effectiveness of the proposed action recognition approach.
提出了一种新的视觉无关的人体动作识别方法。多摄像头装置用于从不同视角捕捉人体。动作通过一种新的动作表示来描述,即所谓的多视图动作图像(MVAI),它有效地解决了摄像机视点识别问题,即识别每个摄像机相对于人的身体的位置。将线性判别分析应用于mvai,将动作映射到判别特征空间,在判别特征空间中使用简单的最近类质心分类方案对动作进行分类。实验结果表明了所提出的动作识别方法的有效性。
{"title":"View-independent human action recognition based on multi-view action images and discriminant learning","authors":"Alexandros Iosifidis, A. Tefas, I. Pitas","doi":"10.1109/IVMSPW.2013.6611931","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611931","url":null,"abstract":"In this paper a novel view-independent human action recognition method is proposed. A multi-camera setup is used to capture the human body from different viewing angles. Actions are described by a novel action representation, the so-called multi-view action image (MVAI), which effectively addresses the camera viewpoint identification problem, i.e., the identification of the position of each camera with respect to the person's body. Linear Discriminant Analysis is applied on the MVAIs in order to to map actions to a discriminant feature space where actions are classified by using a simple nearest class centroid classification scheme. Experimental results denote the effectiveness of the proposed action recognition approach.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134087718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
3-D video depth map quantization based on Lloyd's algorithm 基于Lloyd算法的三维视频深度图量化
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611903
Zhi Jin, T. Tillo, Jimin Xiao, Fei Cheng
Depth map based 3-D view synthesis technique allows to synthesize different virtual perspectives of the same 3-D content, and the lifelike 3-D view perception of these virtual perspectives is highly depending on the accuracy of depth information. So, although depth map is not viewed by the users, however, its quality directly affects that of the synthesized virtual views by the 3-D warping process. In view of current progress in depth map acquisition techniques, and in order to maintain compatibility with different encoding standards depth values need to undergo quantization, so as to be represented at practical and convenient values. However, the quantization process causes depth map distortion. In order to reduce this negative effect, this paper propose a new quantization method which jointly considers the distribution of depth values and the 3-D rendering process, so as to minimize the depth map induced rendering distortion. The experimental results with the proposed quantization method indicate that the SNR gain has increased from 1dB to 14dB for different sequences when compared with the traditional non-uniform quantization method.
基于深度图的三维视图合成技术允许对相同的三维内容合成不同的虚拟视角,而这些虚拟视角的逼真三维视图感知高度依赖于深度信息的准确性。因此,纵深图虽然不是用户看到的,但其质量直接影响到三维变形合成的虚拟视图的质量。鉴于目前深度图采集技术的进展,为了保持与不同编码标准的兼容性,需要对深度值进行量化,以便表示为实用方便的值。然而,量化过程会导致深度图失真。为了减少这种负面影响,本文提出了一种新的量化方法,该方法综合考虑了深度值的分布和三维渲染过程,从而最大限度地减少了深度图引起的渲染失真。实验结果表明,与传统的非均匀量化方法相比,该量化方法对不同序列的信噪比增益从1dB提高到14dB。
{"title":"3-D video depth map quantization based on Lloyd's algorithm","authors":"Zhi Jin, T. Tillo, Jimin Xiao, Fei Cheng","doi":"10.1109/IVMSPW.2013.6611903","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611903","url":null,"abstract":"Depth map based 3-D view synthesis technique allows to synthesize different virtual perspectives of the same 3-D content, and the lifelike 3-D view perception of these virtual perspectives is highly depending on the accuracy of depth information. So, although depth map is not viewed by the users, however, its quality directly affects that of the synthesized virtual views by the 3-D warping process. In view of current progress in depth map acquisition techniques, and in order to maintain compatibility with different encoding standards depth values need to undergo quantization, so as to be represented at practical and convenient values. However, the quantization process causes depth map distortion. In order to reduce this negative effect, this paper propose a new quantization method which jointly considers the distribution of depth values and the 3-D rendering process, so as to minimize the depth map induced rendering distortion. The experimental results with the proposed quantization method indicate that the SNR gain has increased from 1dB to 14dB for different sequences when compared with the traditional non-uniform quantization method.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116218544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Frame rate up conversion of 3D video by motion and depth fusion 基于运动与深度融合的3D视频帧率提升转换
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611897
Yeejin Lee, Zucheul Lee, Truong Q. Nguyen
This paper presents a new frame-rate up conversion method for video plus depth representation. The proposed method improves the quality of video using additional depth information. The depth information is utilized in three ways to improve the accuracy of motion estimation. One usage is that depth frame is added in block matching criterion to estimate initial motion. Another usage is to detect object boundary region combining depth value variation and local depth distribution adjustment technique. Finally, motion vector field is refined adaptively by segmenting macroblocks into object regions. Any existing block-based motion-compensated frame interpolation methods may utilize our method to refine motion vector field by using depth frames. Experimental results verify that the proposed method outperforms the conventional motion-compensated frame interpolation algorithms, while preserving object structure.
提出了一种新的视频加深度表示的帧率上升转换方法。该方法利用附加的深度信息提高了视频的质量。通过三种方式利用深度信息来提高运动估计的精度。一种应用是在块匹配准则中加入深度帧来估计初始运动。另一种方法是结合深度值变化和局部深度分布平差技术检测目标边界区域。最后,通过将宏块分割成目标区域,自适应细化运动向量场。任何现有的基于块的运动补偿帧插值方法都可以利用我们的方法通过使用深度帧来细化运动矢量场。实验结果表明,该方法在保留目标结构的前提下,优于传统的运动补偿帧插值算法。
{"title":"Frame rate up conversion of 3D video by motion and depth fusion","authors":"Yeejin Lee, Zucheul Lee, Truong Q. Nguyen","doi":"10.1109/IVMSPW.2013.6611897","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611897","url":null,"abstract":"This paper presents a new frame-rate up conversion method for video plus depth representation. The proposed method improves the quality of video using additional depth information. The depth information is utilized in three ways to improve the accuracy of motion estimation. One usage is that depth frame is added in block matching criterion to estimate initial motion. Another usage is to detect object boundary region combining depth value variation and local depth distribution adjustment technique. Finally, motion vector field is refined adaptively by segmenting macroblocks into object regions. Any existing block-based motion-compensated frame interpolation methods may utilize our method to refine motion vector field by using depth frames. Experimental results verify that the proposed method outperforms the conventional motion-compensated frame interpolation algorithms, while preserving object structure.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134404780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
3D image quality index using SDP-based binocular perception model 基于sdp双目感知模型的三维图像质量指数
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611936
Hyunsuk Ko, Chang-Su Kim, S. Choi, C.-C. Jay Kuo
A novel quality index for stereoscopic image pairs is proposed in this work. First, we introduce a parameter called the structural distortion parameter (SDP), which varies according to different distortion types. Then, we use the SDP as a control parameter in a binocular perception model, and apply it into three components of SSIM to obtain an overall quality index. In the proposed framework, the binocular model accommodates distortion types and distortion degrees and, therefore, offers robust quality assessment results for both symmetric and asymmetric distortions. It is shown by experimental results that the proposed index outperforms several existing 3D image quality index methods.
本文提出了一种新的立体像对质量指标。首先,我们引入一个叫做结构畸变参数(SDP)的参数,该参数根据不同的畸变类型而变化。然后,我们将SDP作为双目感知模型的控制参数,并将其应用于SSIM的三个组成部分,以获得整体质量指标。在提出的框架中,双目模型适应畸变类型和畸变程度,因此,对于对称和非对称畸变都提供了可靠的质量评估结果。实验结果表明,该指数优于现有的几种三维图像质量指数方法。
{"title":"3D image quality index using SDP-based binocular perception model","authors":"Hyunsuk Ko, Chang-Su Kim, S. Choi, C.-C. Jay Kuo","doi":"10.1109/IVMSPW.2013.6611936","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611936","url":null,"abstract":"A novel quality index for stereoscopic image pairs is proposed in this work. First, we introduce a parameter called the structural distortion parameter (SDP), which varies according to different distortion types. Then, we use the SDP as a control parameter in a binocular perception model, and apply it into three components of SSIM to obtain an overall quality index. In the proposed framework, the binocular model accommodates distortion types and distortion degrees and, therefore, offers robust quality assessment results for both symmetric and asymmetric distortions. It is shown by experimental results that the proposed index outperforms several existing 3D image quality index methods.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122682266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Towards naturalistic depth propagation 走向自然的深度传播
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611934
Weicheng Huang, Xun Cao, K. Lu, Qionghai Dai, A. Bovik
We propose a two-stage depth propagation algorithm for semi-automatic 2D-to-3D video conversion that forces the solution towards statistical “naturalness”. First, both forward and backward motion vectors are estimated and compared to decide initial depth values, then a compensation process is adopted to further improve the depth initialization. Secondly, the luminance and initial depth are decomposed into a wavelet pyramid. Each sub-band of depth is inferred using a Bayesian formulation under a natural scene statistic prior assumption. This is incorporated into a propagation target function as a prior regularizing term. The final depth map associated with each frame of the input 2D video is optimized by composing all the sub-bands. Experimental results obtained on various sequences show that the presented method outperforms several state-of-the-art depth propagation methods.
我们提出了一种用于半自动2d到3d视频转换的两阶段深度传播算法,该算法迫使解决方案趋向于统计“自然性”。首先对前后运动矢量进行估计和比较,确定初始深度值,然后采用补偿过程进一步改进深度初始化。其次,将亮度和初始深度分解成小波金字塔;每个深度子带在自然场景统计先验假设下使用贝叶斯公式进行推断。这被合并到传播目标函数中作为一个先验正则化项。通过组合所有子带来优化与输入的2D视频的每帧相关的最终深度图。在不同序列上的实验结果表明,该方法优于几种最先进的深度传播方法。
{"title":"Towards naturalistic depth propagation","authors":"Weicheng Huang, Xun Cao, K. Lu, Qionghai Dai, A. Bovik","doi":"10.1109/IVMSPW.2013.6611934","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611934","url":null,"abstract":"We propose a two-stage depth propagation algorithm for semi-automatic 2D-to-3D video conversion that forces the solution towards statistical “naturalness”. First, both forward and backward motion vectors are estimated and compared to decide initial depth values, then a compensation process is adopted to further improve the depth initialization. Secondly, the luminance and initial depth are decomposed into a wavelet pyramid. Each sub-band of depth is inferred using a Bayesian formulation under a natural scene statistic prior assumption. This is incorporated into a propagation target function as a prior regularizing term. The final depth map associated with each frame of the input 2D video is optimized by composing all the sub-bands. Experimental results obtained on various sequences show that the presented method outperforms several state-of-the-art depth propagation methods.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130688454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A novel one shot object profilometry system using Direct Sequence Spread Spectrum profilometry 采用直接序列扩频轮廓术的一种新的单镜头物体轮廓术系统
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611896
S. Woolford, I. Burnett
In this paper a new method of determining 3D object shape using patterns derived from Direct Sequence Spread Spectrum (DSSS) and an Unscented Kalman Filter (UKF) is presented. First a binary message is encoded via Binary Phase Shift Keying (BPSK), and spread using pseudo-random spreading to create a pattern. An Iterative Unscented Kalman Filter (IUKF) is then used to determine the deformation in the pattern due to an object, and a Kalman smoother is used to reduce noise in the deformation estimation. Results show that the iterative UKF is able to determine the deformation in the pattern with a lower absolute error residual between the ground truth and estimated deformation than the non-iterated UKF. Results of the accompanying Cramer-Rao lower bounds show that the lower bound on the DSSS Pattern is lower than that of the fringe pattern.
本文提出了一种利用直接序列扩频(DSSS)和无气味卡尔曼滤波(UKF)得到的模式来确定三维物体形状的新方法。首先,通过二进制相移键控(BPSK)对二进制消息进行编码,并使用伪随机传播来创建模式。然后使用迭代无气味卡尔曼滤波器(IUKF)来确定图案中由于物体引起的变形,并使用卡尔曼平滑来降低变形估计中的噪声。结果表明,与非迭代UKF相比,迭代UKF能够确定模式中的变形,并且地面真值与估计变形之间的绝对残差更小。伴随的Cramer-Rao下界结果表明,DSSS图的下界低于条纹图的下界。
{"title":"A novel one shot object profilometry system using Direct Sequence Spread Spectrum profilometry","authors":"S. Woolford, I. Burnett","doi":"10.1109/IVMSPW.2013.6611896","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611896","url":null,"abstract":"In this paper a new method of determining 3D object shape using patterns derived from Direct Sequence Spread Spectrum (DSSS) and an Unscented Kalman Filter (UKF) is presented. First a binary message is encoded via Binary Phase Shift Keying (BPSK), and spread using pseudo-random spreading to create a pattern. An Iterative Unscented Kalman Filter (IUKF) is then used to determine the deformation in the pattern due to an object, and a Kalman smoother is used to reduce noise in the deformation estimation. Results show that the iterative UKF is able to determine the deformation in the pattern with a lower absolute error residual between the ground truth and estimated deformation than the non-iterated UKF. Results of the accompanying Cramer-Rao lower bounds show that the lower bound on the DSSS Pattern is lower than that of the fringe pattern.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"201 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121618541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved indoor scene geometry recognition from single image based on depth map 改进了基于深度图的单幅室内场景几何识别方法
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611938
Yixian Liu, Xinyu Lin, Qianni Zhang, E. Izquierdo
Interpreting 3D structure from 2D images is a constant problem to be solved in the field of computer vision. Prior work has been made to tackle this issue mainly in two different ways - depth estimation from multiple-view images based on geometric triangulation and depth reasoning from single image depending on monocular depth cues. Both solutions do not involve direct depth map information. In this work, we captured a RGBD dataset using Microsoft Kinect depth sensor. Approximate depth information is acquired as the fourth channel and employed as an extra reference for 3D scene geometry reasoning. It helps to achieve better estimation accuracy. We define nine basic geometric models for general indoor restricted-view scenes. Then we extract low/medium level colour and depth features from all four of the RGBD channels. Sequential Minimal Optimization SVM is used in this work as efficient classification tool. Experiments are implemented to compare the result of this approach with previous work that does not have the depth channel as input.
从二维图像中解释三维结构是计算机视觉领域一直需要解决的问题。先前的工作主要通过两种不同的方式来解决这个问题——基于几何三角剖分的多视图图像深度估计和基于单目深度线索的单幅图像深度推理。这两种解决方案都不涉及直接的深度图信息。在这项工作中,我们使用微软Kinect深度传感器捕获了RGBD数据集。获取近似深度信息作为第四通道,作为三维场景几何推理的额外参考。它有助于实现更好的估计精度。我们定义了9个基本的几何模型,用于一般的室内受限视图场景。然后我们从所有四个RGBD通道中提取低/中等水平的颜色和深度特征。本文采用了序列最小优化支持向量机作为高效的分类工具。通过实验将这种方法的结果与以前没有深度通道作为输入的工作进行比较。
{"title":"Improved indoor scene geometry recognition from single image based on depth map","authors":"Yixian Liu, Xinyu Lin, Qianni Zhang, E. Izquierdo","doi":"10.1109/IVMSPW.2013.6611938","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611938","url":null,"abstract":"Interpreting 3D structure from 2D images is a constant problem to be solved in the field of computer vision. Prior work has been made to tackle this issue mainly in two different ways - depth estimation from multiple-view images based on geometric triangulation and depth reasoning from single image depending on monocular depth cues. Both solutions do not involve direct depth map information. In this work, we captured a RGBD dataset using Microsoft Kinect depth sensor. Approximate depth information is acquired as the fourth channel and employed as an extra reference for 3D scene geometry reasoning. It helps to achieve better estimation accuracy. We define nine basic geometric models for general indoor restricted-view scenes. Then we extract low/medium level colour and depth features from all four of the RGBD channels. Sequential Minimal Optimization SVM is used in this work as efficient classification tool. Experiments are implemented to compare the result of this approach with previous work that does not have the depth channel as input.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"308 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129750645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Shape From Focus by Total Variation 形状从焦点的总变化
Pub Date : 2013-06-10 DOI: 10.1109/IVMSPW.2013.6611940
M. Mahmood
Usually, in Shape From Focus (SFF) methods, a single focus measure is applied on an image sequence to obtain an initial depth map and then an approximation technique is used to recover three-dimensional (3D) shape of an object. However, different focus measures perform differently in diverse conditions and it is hard to get accurate 3D shape based on a single focus measure. In this paper, we propose a nonlinear Total Variation (TV) based method for recovering 3D shape of an object by diffusing several initial depth maps obtained through different focus measures. Several experiments have been conducted using images of synthetic and real objects to evaluate the performance of the proposed method. Comparative analysis demonstrates of the effectiveness of the proposed approach.
形状从焦点(Shape From Focus, SFF)方法通常是对图像序列进行单焦点测量获得初始深度图,然后使用近似技术恢复物体的三维形状。然而,不同的焦距测量在不同的条件下表现不同,基于单一的焦距测量很难获得准确的三维形状。本文提出了一种基于非线性全变分(TV)的方法,通过对不同焦点测量得到的多个初始深度图进行扩散来恢复物体的三维形状。利用合成物体和真实物体的图像进行了多次实验,以评估所提出方法的性能。对比分析表明了该方法的有效性。
{"title":"Shape From Focus by Total Variation","authors":"M. Mahmood","doi":"10.1109/IVMSPW.2013.6611940","DOIUrl":"https://doi.org/10.1109/IVMSPW.2013.6611940","url":null,"abstract":"Usually, in Shape From Focus (SFF) methods, a single focus measure is applied on an image sequence to obtain an initial depth map and then an approximation technique is used to recover three-dimensional (3D) shape of an object. However, different focus measures perform differently in diverse conditions and it is hard to get accurate 3D shape based on a single focus measure. In this paper, we propose a nonlinear Total Variation (TV) based method for recovering 3D shape of an object by diffusing several initial depth maps obtained through different focus measures. Several experiments have been conducted using images of synthetic and real objects to evaluate the performance of the proposed method. Comparative analysis demonstrates of the effectiveness of the proposed approach.","PeriodicalId":170714,"journal":{"name":"IVMSP 2013","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133214272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
期刊
IVMSP 2013
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1