首页 > 最新文献

2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)最新文献

英文 中文
FAST OCCLUSION FILLING METHOD FOR MULTIVIEW VIDEO GENERATION 多视点视频生成的快速遮挡填充方法
A. Khatiullin, Mikhail Erofeev, D. Vatolin
Occlusion filling is a basic problem for multiview video generation from existing monocular video. The essential goal of this problem is to recover missing information about a scenes 3D structure and corresponding texture.We propose a method for content-aware deformation of the source view that ensures no disoccluded regions are visible in the synthesized views while also keeping visible distortions to a minimum. We formulate this problem in terms of global energy min-imization. Furthermore, we introduce a similar variable-rejection algorithm that, along with other known optimization techniques, allows us to accelerate the energy function minimization by nearly 30 times and still maintain the visual quality of the synthesized views.
遮挡填充是现有单目视频生成多视点视频的一个基本问题。该问题的基本目标是恢复场景三维结构和相应纹理的缺失信息。我们提出了一种源视图的内容感知变形方法,该方法确保在合成视图中看不到未遮挡的区域,同时将可见的变形保持在最低限度。我们用全球能源最小化来表述这个问题。此外,我们引入了一个类似的变量拒绝算法,该算法与其他已知的优化技术一起,使我们能够将能量函数最小化的速度提高近30倍,并且仍然保持合成视图的视觉质量。
{"title":"FAST OCCLUSION FILLING METHOD FOR MULTIVIEW VIDEO GENERATION","authors":"A. Khatiullin, Mikhail Erofeev, D. Vatolin","doi":"10.1109/3DTV.2018.8478562","DOIUrl":"https://doi.org/10.1109/3DTV.2018.8478562","url":null,"abstract":"Occlusion filling is a basic problem for multiview video generation from existing monocular video. The essential goal of this problem is to recover missing information about a scenes 3D structure and corresponding texture.We propose a method for content-aware deformation of the source view that ensures no disoccluded regions are visible in the synthesized views while also keeping visible distortions to a minimum. We formulate this problem in terms of global energy min-imization. Furthermore, we introduce a similar variable-rejection algorithm that, along with other known optimization techniques, allows us to accelerate the energy function minimization by nearly 30 times and still maintain the visual quality of the synthesized views.","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"134 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116336615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
DESIGN OF MICRO PHOTON SIEVE ARRAYS FOR HIGH RESOLUTION LIGHT-FIELD CAPTURE IN PLENOPTIC CAMERAS 用于全光学相机高分辨率光场捕获的微光子筛阵列设计
Ali Özgür Yöntem, D. Chu
The design of micro photon sieve arrays (PSAs) is investigated for light-field capture with high spatial resolution in plenoptic cameras. A commercial very high-resolution full-frame camera with a manual lens is converted into a plenoptic camera for high-resolution depth image acquisition by using the designed PSA as an add-on diffractive optical element in place of an ordinary refractive microlens array or a diffractive micro Fresnel Zone Plate (FZP) array, which is used in integral imaging applications. The noise introduced by the diffractive nature of the optical element is reduced by standard image processing tools. The light-field data is also used for computational refocusing of the 3D scene with wave propagation tools.
研究了用于全光学相机高空间分辨率光场捕获的微光子筛阵列的设计。将设计的PSA作为附加的衍射光学元件,取代用于集成成像应用的普通折射微透镜阵列或衍射微菲涅耳带板阵列,将带有手动镜头的商用高分辨率全画幅相机转换为高分辨率深度图像采集的全光学相机。由光学元件的衍射特性引入的噪声通过标准图像处理工具减少。光场数据也被用于用波传播工具计算三维场景的重新聚焦。
{"title":"DESIGN OF MICRO PHOTON SIEVE ARRAYS FOR HIGH RESOLUTION LIGHT-FIELD CAPTURE IN PLENOPTIC CAMERAS","authors":"Ali Özgür Yöntem, D. Chu","doi":"10.1109/3DTV.2018.8478587","DOIUrl":"https://doi.org/10.1109/3DTV.2018.8478587","url":null,"abstract":"The design of micro photon sieve arrays (PSAs) is investigated for light-field capture with high spatial resolution in plenoptic cameras. A commercial very high-resolution full-frame camera with a manual lens is converted into a plenoptic camera for high-resolution depth image acquisition by using the designed PSA as an add-on diffractive optical element in place of an ordinary refractive microlens array or a diffractive micro Fresnel Zone Plate (FZP) array, which is used in integral imaging applications. The noise introduced by the diffractive nature of the optical element is reduced by standard image processing tools. The light-field data is also used for computational refocusing of the 3D scene with wave propagation tools.","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130366456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
ICP WITH DEPTH COMPENSATION FOR CALIBRATION OF MULTIPLE TOF SENSORS 具有深度补偿的Icp,用于校准多个tof传感器
Norishige Fukushima
We propose an iterative closest point (ICP) based calibration for time of flight (ToF) multiple depth sensors. For the multiple sensor calibrations, we usually use 2D patterns calibration with IR images. The depth sensor output depends on calibration parameters at a factory; thus, the re-calibration must include gaps from the calibration in the factory. Therefore, we use direct correspondences among depth values, and the calibrating extrinsic parameters by using ICP. Usually, simultaneous localization and mapping (SLAM) uses ICP, such as KinectFusion. The case of multiple sensor calibrations, however, is harder than the SLAM case. In this case, the distance between cameras is too far to apply ICP. Therefore, we modify the ICP based calibration for multiple sensors. The proposed method uses specific calibration objects to enforce the matching ability among sensors. Also, we proposed a compensation method for ToF depth map distortions.
我们提出了一种基于迭代最近点(ICP)的飞行时间(ToF)多深度传感器校准方法。对于多传感器校准,我们通常使用红外图像的二维模式校准。深度传感器的输出取决于工厂的校准参数;因此,重新校准必须包括与工厂校准的差距。因此,我们使用深度值之间的直接对应关系,并使用ICP校准外部参数。通常,同步定位和映射(SLAM)使用ICP,例如KinectFusion。然而,多传感器校准的情况比SLAM的情况更难。在这种情况下,摄像机之间的距离太远,无法应用ICP。因此,我们修改了基于ICP的多传感器校准方法。该方法利用特定的标定对象来增强传感器间的匹配能力。同时,提出了一种补偿ToF深度图失真的方法。
{"title":"ICP WITH DEPTH COMPENSATION FOR CALIBRATION OF MULTIPLE TOF SENSORS","authors":"Norishige Fukushima","doi":"10.1109/3DTV.2018.8478527","DOIUrl":"https://doi.org/10.1109/3DTV.2018.8478527","url":null,"abstract":"We propose an iterative closest point (ICP) based calibration for time of flight (ToF) multiple depth sensors. For the multiple sensor calibrations, we usually use 2D patterns calibration with IR images. The depth sensor output depends on calibration parameters at a factory; thus, the re-calibration must include gaps from the calibration in the factory. Therefore, we use direct correspondences among depth values, and the calibrating extrinsic parameters by using ICP. Usually, simultaneous localization and mapping (SLAM) uses ICP, such as KinectFusion. The case of multiple sensor calibrations, however, is harder than the SLAM case. In this case, the distance between cameras is too far to apply ICP. Therefore, we modify the ICP based calibration for multiple sensors. The proposed method uses specific calibration objects to enforce the matching ability among sensors. Also, we proposed a compensation method for ToF depth map distortions.","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115844791","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
REAL-TIME MULTI-VIEW VOLUMETRIC RECONSTRUCTION OF DYNAMIC SCENES USING KINECT V2 基于kinect v2的动态场景实时多视角体重建
Andrej Satnik, E. Izquierdo
A key challenge when displaying and processing sensed real-time 3D data is efficiency of generating and post-processing algorithms in order to acquire high quality 3D content. In contrast, our approach focuses on volumetric generation and processing volumetric data using an efficient low-cost hardware setting. Acquisition of volumetric data is performed by connecting several Kinect v2 scanners to a single PC that are subsequently calibrated using planar pattern. This process is by no means trivial and requires well designed algorithms for fast processing and quick rendering of volumetric data. This can be achieved by fusing efficient filtering methods such as Weighted median filter (WM), Radius outlier removal (ROR) and Laplace-based smoothing algorithm. In this context, we demonstrate the robustness and efficiency of our technique by sensing several scenes.
为了获得高质量的三维内容,在显示和处理实时三维数据时,一个关键的挑战是生成和后处理算法的效率。相比之下,我们的方法侧重于使用高效低成本的硬件设置生成和处理体积数据。体积数据的获取是通过将多个Kinect v2扫描仪连接到一台PC上进行的,随后使用平面模式进行校准。这个过程绝不是微不足道的,需要设计良好的算法来快速处理和快速渲染体积数据。这可以通过融合有效的滤波方法,如加权中值滤波(WM)、半径离群值去除(ROR)和基于拉普拉斯的平滑算法来实现。在这种情况下,我们通过感知几个场景来证明我们的技术的鲁棒性和效率。
{"title":"REAL-TIME MULTI-VIEW VOLUMETRIC RECONSTRUCTION OF DYNAMIC SCENES USING KINECT V2","authors":"Andrej Satnik, E. Izquierdo","doi":"10.1109/3DTV.2018.8478536","DOIUrl":"https://doi.org/10.1109/3DTV.2018.8478536","url":null,"abstract":"A key challenge when displaying and processing sensed real-time 3D data is efficiency of generating and post-processing algorithms in order to acquire high quality 3D content. In contrast, our approach focuses on volumetric generation and processing volumetric data using an efficient low-cost hardware setting. Acquisition of volumetric data is performed by connecting several Kinect v2 scanners to a single PC that are subsequently calibrated using planar pattern. This process is by no means trivial and requires well designed algorithms for fast processing and quick rendering of volumetric data. This can be achieved by fusing efficient filtering methods such as Weighted median filter (WM), Radius outlier removal (ROR) and Laplace-based smoothing algorithm. In this context, we demonstrate the robustness and efficiency of our technique by sensing several scenes.","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130613019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
DEPTH PERCEPTION PREDICTION OF 3D VIDEO FOR ENSURING ADVANCED MULTIMEDIA SERVICES 3d视频深度感知预测,确保先进的多媒体服务
G. Nur, F. Battisti
A key role in the advancement of 3 Dimensional TV services is played by the development of 3D video quality metrics used for the assessment of the perceived quality. Moreover, this key role can only be supported when the features associated with the 3D video nature is reliably and efficiently characterized in these metrics. In this study, z-direction motion incorporated with significant depth levels in depth map sequences are considered as the main characterizations of the 3D nature. The 3D video quality metrics can be classified into three categories based on the need for the reference video during the assessment process at the user end: Full Reference (FR), Reduced Reference (RR) and No Reference (NR). In this study we propose a NR quality metric, PNRM, suitable for on-the-fly 3D video services. In order to evaluate the reliability and effectiveness of the proposed metric, subjective experiments are conducted in this paper. Observing the high correlation with the subjective experimental results, it can be clearly stated that the proposed metric is able to mimic the Human Visual System (HVS).
用于评估感知质量的3D视频质量指标的开发在三维电视服务的发展中起着关键作用。此外,只有当与3D视频性质相关的特征在这些指标中得到可靠和有效的表征时,才能支持这一关键作用。在本研究中,深度图序列中包含显著深度层次的z方向运动被认为是三维性质的主要表征。根据用户在评估过程中对参考视频的需求,可以将3D视频质量指标分为三类:完整参考(Full reference)、减少参考(Reduced reference)和无参考(No reference)。在本研究中,我们提出了一种适合于实时3D视频服务的NR质量度量,PNRM。为了评估所提出的度量的可靠性和有效性,本文进行了主观实验。观察到与主观实验结果的高度相关性,可以清楚地说明所提出的度量能够模拟人类视觉系统(HVS)。
{"title":"DEPTH PERCEPTION PREDICTION OF 3D VIDEO FOR ENSURING ADVANCED MULTIMEDIA SERVICES","authors":"G. Nur, F. Battisti","doi":"10.1109/3DTV.2018.8478491","DOIUrl":"https://doi.org/10.1109/3DTV.2018.8478491","url":null,"abstract":"A key role in the advancement of 3 Dimensional TV services is played by the development of 3D video quality metrics used for the assessment of the perceived quality. Moreover, this key role can only be supported when the features associated with the 3D video nature is reliably and efficiently characterized in these metrics. In this study, z-direction motion incorporated with significant depth levels in depth map sequences are considered as the main characterizations of the 3D nature. The 3D video quality metrics can be classified into three categories based on the need for the reference video during the assessment process at the user end: Full Reference (FR), Reduced Reference (RR) and No Reference (NR). In this study we propose a NR quality metric, PNRM, suitable for on-the-fly 3D video services. In order to evaluate the reliability and effectiveness of the proposed metric, subjective experiments are conducted in this paper. Observing the high correlation with the subjective experimental results, it can be clearly stated that the proposed metric is able to mimic the Human Visual System (HVS).","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130671176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3DTV-CON 2018 Organizing Committee Page 3DTV-CON 2018组委会页面
{"title":"3DTV-CON 2018 Organizing Committee Page","authors":"","doi":"10.1109/3dtv.2018.8478442","DOIUrl":"https://doi.org/10.1109/3dtv.2018.8478442","url":null,"abstract":"","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129172257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SIGN LANGUAGE RECOGNITION BASED ON HAND AND BODY SKELETAL DATA 基于手和身体骨骼数据的手语识别
D. Konstantinidis, K. Dimitropoulos, P. Daras
Sign language recognition (SLR) is a challenging, but highly important research field for several computer vision systems that attempt to facilitate the communication among the deaf and hearing impaired people. In this work, we propose an accurate and robust deep learning-based methodology for sign language recognition from video sequences. Our novel method relies on hand and body skeletal features extracted from RGB videos and, therefore, it acquires highly discriminative for gesture recognition skeletal data without the need for any additional equipment, such as data gloves, that may restrict signer’s movements. Experimentation on a large publicly available sign language dataset reveals the superiority of our methodology with respect to other state of the art approaches relying solely on RGB features.
手语识别(SLR)是一个具有挑战性的研究领域,但对于一些试图促进聋人和听障人士之间交流的计算机视觉系统来说,这是一个非常重要的研究领域。在这项工作中,我们提出了一种准确而稳健的基于深度学习的方法,用于从视频序列中识别手语。我们的新方法依赖于从RGB视频中提取的手和身体骨骼特征,因此,它获得了高度判别的手势识别骨骼数据,而不需要任何额外的设备,如数据手套,这可能会限制签署人的运动。在一个大型公开可用的手语数据集上的实验揭示了我们的方法相对于仅依赖RGB特征的其他最先进方法的优越性。
{"title":"SIGN LANGUAGE RECOGNITION BASED ON HAND AND BODY SKELETAL DATA","authors":"D. Konstantinidis, K. Dimitropoulos, P. Daras","doi":"10.1109/3DTV.2018.8478467","DOIUrl":"https://doi.org/10.1109/3DTV.2018.8478467","url":null,"abstract":"Sign language recognition (SLR) is a challenging, but highly important research field for several computer vision systems that attempt to facilitate the communication among the deaf and hearing impaired people. In this work, we propose an accurate and robust deep learning-based methodology for sign language recognition from video sequences. Our novel method relies on hand and body skeletal features extracted from RGB videos and, therefore, it acquires highly discriminative for gesture recognition skeletal data without the need for any additional equipment, such as data gloves, that may restrict signer’s movements. Experimentation on a large publicly available sign language dataset reveals the superiority of our methodology with respect to other state of the art approaches relying solely on RGB features.","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122593895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 61
EXPERT EVALUATION OF A NOVEL LIGHT-FIELD VISUALIZATION FORMAT 一种新型光场可视化格式的专家评价
A. Cserkaszky, P. A. Kara, A. Barsi, M. Martini
Light-field visualization is continuously emerging in industrial sectors, and the appearance on the consumer market is approaching. Yet this process is halted, or at least slowed down, by the lack of proper display-independent light-field formats. Such formats are necessary to enable the efficient interchange between light-field content creation and visualization, and thus support potential future use case scenarios of this technology. In this paper, we introduce the results of a perceived quality assessment research, performed on our own novel light-field visualization format. The subjective tests, which compared conventional linear camera array visualization to our format, were completed by experts only, thus quality assessment was an expert evaluation. We aim to use the findings gathered in this research to carry out a large-scale subjective test series in the future, with non-expert observers.
光场可视化技术在工业领域不断涌现,在消费市场上出现的日子越来越近。然而,由于缺乏适当的与显示无关的光场格式,这一过程被停止了,或者至少放慢了速度。这样的格式对于实现光场内容创建和可视化之间的有效交换是必要的,因此支持该技术的潜在未来用例场景。在本文中,我们介绍了一项感知质量评估研究的结果,该研究是在我们自己的新型光场可视化格式上进行的。将传统线性相机阵列可视化与我们的格式进行比较的主观测试仅由专家完成,因此质量评估是专家评估。我们的目标是利用本研究中收集的结果,在未来进行大规模的主观测试系列,与非专家观察者。
{"title":"EXPERT EVALUATION OF A NOVEL LIGHT-FIELD VISUALIZATION FORMAT","authors":"A. Cserkaszky, P. A. Kara, A. Barsi, M. Martini","doi":"10.1109/3DTV.2018.8478436","DOIUrl":"https://doi.org/10.1109/3DTV.2018.8478436","url":null,"abstract":"Light-field visualization is continuously emerging in industrial sectors, and the appearance on the consumer market is approaching. Yet this process is halted, or at least slowed down, by the lack of proper display-independent light-field formats. Such formats are necessary to enable the efficient interchange between light-field content creation and visualization, and thus support potential future use case scenarios of this technology. In this paper, we introduce the results of a perceived quality assessment research, performed on our own novel light-field visualization format. The subjective tests, which compared conventional linear camera array visualization to our format, were completed by experts only, thus quality assessment was an expert evaluation. We aim to use the findings gathered in this research to carry out a large-scale subjective test series in the future, with non-expert observers.","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124256897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
3DTV-CON 2018 Index
{"title":"3DTV-CON 2018 Index","authors":"","doi":"10.1109/3dtv.2018.8478548","DOIUrl":"https://doi.org/10.1109/3dtv.2018.8478548","url":null,"abstract":"","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"638 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116084692","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ACCURATE METHOD OF TEMPORAL-SHIFT ESTIMATION FOR 3D VIDEO 三维视频时移估计的精确方法
Aleksandr Ploshkin, D. Vatolin
Video synchronization is a fundamental computer-vision task that is necessary for a wide range of applications. A 3D video involves two streams, which show the scene from different angles concurrently, but many cases exhibit desynchronization between them. This paper investigates the problem of synchronizing the left and right stereoscopic views. We assume the temporal shift (time difference) and geometric distortion between the two streams are constant throughout each scene. We propose a temporal-shift estimation method with subframe accuracy based on a block-matching algorithm.
视频同步是一项基本的计算机视觉任务,是广泛应用所必需的。3D视频包含两个流,它们同时从不同的角度显示场景,但在许多情况下它们之间表现出不同步。本文研究了左右立体视图的同步问题。我们假设两个流之间的时间偏移(时间差)和几何失真在每个场景中都是恒定的。提出了一种基于分块匹配算法的子帧精度时移估计方法。
{"title":"ACCURATE METHOD OF TEMPORAL-SHIFT ESTIMATION FOR 3D VIDEO","authors":"Aleksandr Ploshkin, D. Vatolin","doi":"10.1109/3DTV.2018.8478431","DOIUrl":"https://doi.org/10.1109/3DTV.2018.8478431","url":null,"abstract":"Video synchronization is a fundamental computer-vision task that is necessary for a wide range of applications. A 3D video involves two streams, which show the scene from different angles concurrently, but many cases exhibit desynchronization between them. This paper investigates the problem of synchronizing the left and right stereoscopic views. We assume the temporal shift (time difference) and geometric distortion between the two streams are constant throughout each scene. We propose a temporal-shift estimation method with subframe accuracy based on a block-matching algorithm.","PeriodicalId":267389,"journal":{"name":"2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122707096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2018 - 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1