Jing Li, Zhichao Lu, Gang Zeng, Rui Gan, Long Wang, H. Zha
{"title":"A Joint Learning-Based Method for Multi-view Depth Map Super Resolution","authors":"Jing Li, Zhichao Lu, Gang Zeng, Rui Gan, Long Wang, H. Zha","doi":"10.1109/ACPR.2013.89","DOIUrl":null,"url":null,"abstract":"Depth map super resolution from multi-view depth or color images has long been explored. Multi-view stereo methods produce fine details at texture areas, and depth recordings would compensate when stereo doesn't work, e.g. at non-texture regions. However, resolution of depth maps from depth sensors are rather low. Our objective is to produce a high-res depth map by fusing different sensors from multiple views. In this paper we present a learning-based method, and infer a high-res depth map from our synthetic database by minimizing the proposed energy. As depth alone is not sufficient to describe geometry of the scene, we use additional features like normal and curvature, which are able to capture high-frequency details of the surface. Our optimization framework explores multi-view depth and color consistency, normal and curvature similarity between low-res input and the database and smoothness constraints on pixel-wise depth-color coherence as well as on patch borders. Experimental results on both synthetic and real data show that our method outperforms state-of-the-art.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"146 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 2nd IAPR Asian Conference on Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACPR.2013.89","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Depth map super resolution from multi-view depth or color images has long been explored. Multi-view stereo methods produce fine details at texture areas, and depth recordings would compensate when stereo doesn't work, e.g. at non-texture regions. However, resolution of depth maps from depth sensors are rather low. Our objective is to produce a high-res depth map by fusing different sensors from multiple views. In this paper we present a learning-based method, and infer a high-res depth map from our synthetic database by minimizing the proposed energy. As depth alone is not sufficient to describe geometry of the scene, we use additional features like normal and curvature, which are able to capture high-frequency details of the surface. Our optimization framework explores multi-view depth and color consistency, normal and curvature similarity between low-res input and the database and smoothness constraints on pixel-wise depth-color coherence as well as on patch borders. Experimental results on both synthetic and real data show that our method outperforms state-of-the-art.