首页 > 最新文献

2014 IEEE Visual Communications and Image Processing Conference最新文献

英文 中文
A cost-efficient hardware architecture of deblocking filter in HEVC HEVC中一种经济高效的去块滤波器硬件结构
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051541
Xin Ye, Dandan Ding, Lu Yu
This paper presents a hardware architecture of deblocking filter (DBF) for High Efficiency Video Coding (HEVC) by jointly considering system throughput and hardware cost. A hybrid pipeline with two processing levels is adopted to improve system performance. With the hybrid pipeline, only one 1-D filter and single-port on-chip SRAM are used. According to the data dependence between neighbouring edges, a shifted 16×16 basic processing unit as well as corresponding filtering order is proposed. It reduces memory cost and makes the DBF friendlier to work in a coding/decoding system. The proposed hardware architecture is synthesized under 0.13um standard CMOS technology and result shows that it consumes 17.6k gates at an operating frequency of 250MHz. Consequently, the design can support real-time processing of QFHD (3840×2160) video applications at 60 fps.
综合考虑系统吞吐量和硬件成本,提出了一种用于高效视频编码(HEVC)的去块滤波器(DBF)硬件架构。采用两级处理的混合流水线,提高了系统性能。使用混合管道,只使用一个一维滤波器和单端口片上SRAM。根据相邻边之间的数据依赖性,提出了移位的16×16基本处理单元及相应的滤波顺序。它降低了内存成本,并使DBF更易于在编码/解码系统中工作。在0.13um标准CMOS技术下合成了所提出的硬件架构,结果表明,在工作频率为250MHz时,它消耗了17.6k个门。因此,该设计可以支持实时处理60 fps的QFHD (3840×2160)视频应用。
{"title":"A cost-efficient hardware architecture of deblocking filter in HEVC","authors":"Xin Ye, Dandan Ding, Lu Yu","doi":"10.1109/VCIP.2014.7051541","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051541","url":null,"abstract":"This paper presents a hardware architecture of deblocking filter (DBF) for High Efficiency Video Coding (HEVC) by jointly considering system throughput and hardware cost. A hybrid pipeline with two processing levels is adopted to improve system performance. With the hybrid pipeline, only one 1-D filter and single-port on-chip SRAM are used. According to the data dependence between neighbouring edges, a shifted 16×16 basic processing unit as well as corresponding filtering order is proposed. It reduces memory cost and makes the DBF friendlier to work in a coding/decoding system. The proposed hardware architecture is synthesized under 0.13um standard CMOS technology and result shows that it consumes 17.6k gates at an operating frequency of 250MHz. Consequently, the design can support real-time processing of QFHD (3840×2160) video applications at 60 fps.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131397355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Tagged multi-hypothesis motion compensation scheme for video coding 视频编码的标记多假设运动补偿方案
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051518
Lei Chen, Ronggang Wang, Siwei Ma
Accuracy of prediction block (PB) plays a very important role in improving the coding performance. In this paper, we propose tagged multi-hypothesis motion compensation scheme (TMHMC) for inter frames to improve the accuracy of PB. TMHMC not only makes use of temporal correlation between frames but also the spatial correlation as motion vectors of adjacent blocks are used to derive the PB. For entropy coding process, only one motion vector and a tag indicating which adjacent block is used are coded in bit-stream. Adding TMHMC scheme as an additional mode in MPEG internet video coding (TVC) platform, the bitrate saving is up to 12% at the same objective quality compared with anchor. Average bitrate saving is close to 6% over all test sequences. In addition, we also implement the conventional multi-hypothesis motion compensation (MHMC) scheme. 3% bitrate is further saved on average by TMHMC compared with conventional MHMC.
预测块的准确性对提高编码性能起着非常重要的作用。本文提出了一种帧间带标签的多假设运动补偿方案(TMHMC),以提高帧间运动补偿的精度。TMHMC不仅利用帧间的时间相关性,而且利用相邻块的空间相关性作为运动向量来推导PB。在熵编码过程中,比特流中只编码一个运动向量和标识相邻块的标签。在MPEG网络视频编码(TVC)平台中加入TMHMC作为附加模式,在相同的客观质量下,与锚点相比,比特率可节省12%。在所有测试序列中,平均比特率节省接近6%。此外,我们还实现了传统的多假设运动补偿(MHMC)方案。与传统的MHMC相比,TMHMC平均可节省3%的比特率。
{"title":"Tagged multi-hypothesis motion compensation scheme for video coding","authors":"Lei Chen, Ronggang Wang, Siwei Ma","doi":"10.1109/VCIP.2014.7051518","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051518","url":null,"abstract":"Accuracy of prediction block (PB) plays a very important role in improving the coding performance. In this paper, we propose tagged multi-hypothesis motion compensation scheme (TMHMC) for inter frames to improve the accuracy of PB. TMHMC not only makes use of temporal correlation between frames but also the spatial correlation as motion vectors of adjacent blocks are used to derive the PB. For entropy coding process, only one motion vector and a tag indicating which adjacent block is used are coded in bit-stream. Adding TMHMC scheme as an additional mode in MPEG internet video coding (TVC) platform, the bitrate saving is up to 12% at the same objective quality compared with anchor. Average bitrate saving is close to 6% over all test sequences. In addition, we also implement the conventional multi-hypothesis motion compensation (MHMC) scheme. 3% bitrate is further saved on average by TMHMC compared with conventional MHMC.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127482158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A fast coding algorithm based on inter-view correlations for 3D-HEVC 基于视点间相关性的3D-HEVC快速编码算法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051584
Guangsheng Chi, Xin Jin, Qionghai Dai
The newly published 3D-HEVC has received a remarkable response due to its high compression efficiency which is based on High Efficiency Video Coding (HEVC). However, the complexity of its encoding process is also large as a result of introducing the coding units (CU) size decision process together with the rate distortion optimization (RDO) process. In this paper, a fast coding algorithm making good use of the interview correlations is proposed. With the inter-view correlation statistical analysis, the CU depth candidates of the dependent views can be predicted from the independent view instead of the brute force RDO process in determining CU depth. The experimental results show that the proposed method saves 51% time in texture coding and the loss is negligible.
新发布的3D-HEVC由于其基于高效视频编码(high efficiency Video Coding, HEVC)的高压缩效率而获得了显著的反响。然而,由于引入了编码单元(CU)大小决定过程和率失真优化(RDO)过程,其编码过程的复杂性也很大。本文提出了一种充分利用访谈相关性的快速编码算法。通过视图间相关统计分析,可以从独立视图中预测依赖视图的候选CU深度,而不是通过蛮力RDO过程来确定CU深度。实验结果表明,该方法可节省51%的纹理编码时间,且纹理编码损失可以忽略不计。
{"title":"A fast coding algorithm based on inter-view correlations for 3D-HEVC","authors":"Guangsheng Chi, Xin Jin, Qionghai Dai","doi":"10.1109/VCIP.2014.7051584","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051584","url":null,"abstract":"The newly published 3D-HEVC has received a remarkable response due to its high compression efficiency which is based on High Efficiency Video Coding (HEVC). However, the complexity of its encoding process is also large as a result of introducing the coding units (CU) size decision process together with the rate distortion optimization (RDO) process. In this paper, a fast coding algorithm making good use of the interview correlations is proposed. With the inter-view correlation statistical analysis, the CU depth candidates of the dependent views can be predicted from the independent view instead of the brute force RDO process in determining CU depth. The experimental results show that the proposed method saves 51% time in texture coding and the loss is negligible.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114203099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Depth inference with convolutional neural network 卷积神经网络深度推理
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051531
Hu Tian, Bojin Zhuang, Yan Hua, A. Cai
The goal of depth inference from a single image is to assign a depth to each pixel in the image according to the image content. In this paper, we propose a deep learning model for this task. This model consists of a convolutional neural network (CNN) with a linear regressor being as the last layer. The network is trained with raw RGB image patches cropped by a large window centered at each pixel of an image to extract feature representations. Then the depth map of a test image can be efficiently obtained by forward-passing the image through the trained model plus a simple up-sampling. Contrary to most previous methods based on graphical model and depth sampling, our method alleviates the needs for engineered features and for assumptions about semantic information of the scene. We achieve state-of-the-art results on Make 3D dataset, while keeping low computational time at the test time.
单幅图像深度推断的目标是根据图像内容为图像中的每个像素分配深度。在本文中,我们为这项任务提出了一个深度学习模型。该模型由卷积神经网络(CNN)组成,最后一层是线性回归器。该网络使用以图像每个像素为中心的大窗口裁剪的原始RGB图像补丁进行训练,以提取特征表示。然后,通过训练好的模型对图像进行前向传递,再进行简单的上采样,即可有效地得到测试图像的深度图。与以往大多数基于图形模型和深度采样的方法相反,我们的方法减轻了对工程特征和场景语义信息假设的需求。我们在Make 3D数据集上取得了最先进的结果,同时在测试时保持了较低的计算时间。
{"title":"Depth inference with convolutional neural network","authors":"Hu Tian, Bojin Zhuang, Yan Hua, A. Cai","doi":"10.1109/VCIP.2014.7051531","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051531","url":null,"abstract":"The goal of depth inference from a single image is to assign a depth to each pixel in the image according to the image content. In this paper, we propose a deep learning model for this task. This model consists of a convolutional neural network (CNN) with a linear regressor being as the last layer. The network is trained with raw RGB image patches cropped by a large window centered at each pixel of an image to extract feature representations. Then the depth map of a test image can be efficiently obtained by forward-passing the image through the trained model plus a simple up-sampling. Contrary to most previous methods based on graphical model and depth sampling, our method alleviates the needs for engineered features and for assumptions about semantic information of the scene. We achieve state-of-the-art results on Make 3D dataset, while keeping low computational time at the test time.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123853916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Fast intra mode decision for HEVC based on texture characteristic from RMD and MPM 基于RMD和MPM纹理特征的HEVC快速模式内决策
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051618
Dongdong Zhang, Youwei Chen, E. Izquierdo
In this paper, we proposed a fast intra mode decision algorithm for HEVC to further reduce the candidate modes of RDOQ or even skip RDOQ for a PU, which exploited not only the texture consistency of neighbouring PUs reflected by the relation between the optimal RMD mode and MPM, but also the texture characteristic in a PU reflected by the best two RMD modes. Experimental results show that our proposed algorithm can reduce 31.3% intra coding time with 0.56% BD-rate loss on average.
为了进一步减少PU的RDOQ候选模式甚至跳过RDOQ,本文提出了一种HEVC快速模式内决策算法,该算法不仅利用了最优RMD模式与MPM之间的关系所反映的相邻PU的纹理一致性,而且利用了最优两个RMD模式所反映的PU的纹理特性。实验结果表明,该算法平均减少了31.3%的编码时间,BD-rate损失为0.56%。
{"title":"Fast intra mode decision for HEVC based on texture characteristic from RMD and MPM","authors":"Dongdong Zhang, Youwei Chen, E. Izquierdo","doi":"10.1109/VCIP.2014.7051618","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051618","url":null,"abstract":"In this paper, we proposed a fast intra mode decision algorithm for HEVC to further reduce the candidate modes of RDOQ or even skip RDOQ for a PU, which exploited not only the texture consistency of neighbouring PUs reflected by the relation between the optimal RMD mode and MPM, but also the texture characteristic in a PU reflected by the best two RMD modes. Experimental results show that our proposed algorithm can reduce 31.3% intra coding time with 0.56% BD-rate loss on average.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121629845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Hybrid modeling of natural image in wavelet domain 小波域自然图像的混合建模
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051501
Chongwu Tang, Xiaokang Yang, Guangtao Zhai
Natural image is characterized by its highly kurtotic and heavy-tailed distribution in wavelet domain. These typical non-Gaussian statistics are commonly described by generalized Gaussian density (GGD) or α-stable distribution. However, each of the two models has its own deficiency to capture the variety and complexity of real world scenes. Considering the statistical properties of GGD and α-stable distributions respectively, in this paper we propose a hybrid statistical model of natural image's wavelet coefficients which is better in describing the leptokurtosis and heavy tails simultaneously. Based on a linearly weighted fusion of GGD and α-stable functions, we derive the optimal parametric hybrid model, and measure the model accuracy using Kullback-Leibler divergence, which evaluates the similarity between two probability distributions. Experiment results and comparative studies demonstrate that the proposed hybrid model is closer to the true distribution of natural image's wavelet coefficients than single GGD or α-stable modeling.
自然图像在小波域具有高峰度和重尾分布的特点。这些典型的非高斯统计量通常用广义高斯密度(GGD)或α稳定分布来描述。然而,在捕捉真实世界场景的多样性和复杂性方面,这两种模型各有其不足之处。考虑到GGD和α-稳定分布的统计性质,本文提出了一种能较好地同时描述细峰态和重尾态的自然图像小波系数混合统计模型。基于GGD和α-稳定函数的线性加权融合,导出了最优参数混合模型,并利用Kullback-Leibler散度来评价两个概率分布之间的相似性来衡量模型的精度。实验结果和对比研究表明,该混合模型比单一的GGD或α-稳定模型更接近自然图像小波系数的真实分布。
{"title":"Hybrid modeling of natural image in wavelet domain","authors":"Chongwu Tang, Xiaokang Yang, Guangtao Zhai","doi":"10.1109/VCIP.2014.7051501","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051501","url":null,"abstract":"Natural image is characterized by its highly kurtotic and heavy-tailed distribution in wavelet domain. These typical non-Gaussian statistics are commonly described by generalized Gaussian density (GGD) or α-stable distribution. However, each of the two models has its own deficiency to capture the variety and complexity of real world scenes. Considering the statistical properties of GGD and α-stable distributions respectively, in this paper we propose a hybrid statistical model of natural image's wavelet coefficients which is better in describing the leptokurtosis and heavy tails simultaneously. Based on a linearly weighted fusion of GGD and α-stable functions, we derive the optimal parametric hybrid model, and measure the model accuracy using Kullback-Leibler divergence, which evaluates the similarity between two probability distributions. Experiment results and comparative studies demonstrate that the proposed hybrid model is closer to the true distribution of natural image's wavelet coefficients than single GGD or α-stable modeling.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121887276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Improved disparity vector derivation for inter-view residual prediction in 3D-HEVC 改进的视差矢量推导用于3D-HEVC视间残差预测
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051508
Shiori Sugimoto, S. Shimizu, Akira Kojima
Inter-view residual prediction and advanced residual prediction (ARP) are efficient tools for coding dependent views of 3D video. It can predict the residue of motion compensated prediction (MCP) using the additional derived disparity vector. However, its coding performance depends on the accuracy of disparity vector derivation. In this paper, we propose the improved disparity vector derivation scheme for ARP. In the proposed scheme, the disparity vector is derived from the corresponding block in the reference picture. And that corresponding block is pointed by the same as MCP. Moreover, the disparity vector can be derived not only from the current reference block, but also from the blocks on the all other reference pictures included in the current reference picture lists. In addition, the disparity vector can be derived from both of the blocks predicted by disparity compensated prediction (DCP) and the blocks predicted by MCP and ARP because the derived disparity vector for ARP is stored in additional disparity vector field. Experimental results show that 0.2% bitrate reduction of synthesized views, and up to 0.5% for each dependent views in the reference software of 3D-HEVC.
视点间残差预测和高级残差预测(ARP)是三维视频相关视点编码的有效工具。它可以利用附加的视差矢量来预测运动补偿预测的残差。然而,其编码性能取决于视差矢量推导的准确性。本文提出了一种改进的视差矢量推导方案。在该方案中,视差矢量从参考图像的相应块中导出。相应的块由相同的MCP指向。此外,视差矢量不仅可以从当前参考块中导出,还可以从当前参考图片列表中包含的所有其他参考图片上的块中导出。此外,视差矢量既可以由视差补偿预测(DCP)预测的块导出,也可以由MCP和ARP预测的块导出,因为ARP导出的视差矢量存储在附加的视差矢量场中。实验结果表明,在3D-HEVC参考软件中,合成视图的比特率降低了0.2%,每个依赖视图的比特率降低了0.5%。
{"title":"Improved disparity vector derivation for inter-view residual prediction in 3D-HEVC","authors":"Shiori Sugimoto, S. Shimizu, Akira Kojima","doi":"10.1109/VCIP.2014.7051508","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051508","url":null,"abstract":"Inter-view residual prediction and advanced residual prediction (ARP) are efficient tools for coding dependent views of 3D video. It can predict the residue of motion compensated prediction (MCP) using the additional derived disparity vector. However, its coding performance depends on the accuracy of disparity vector derivation. In this paper, we propose the improved disparity vector derivation scheme for ARP. In the proposed scheme, the disparity vector is derived from the corresponding block in the reference picture. And that corresponding block is pointed by the same as MCP. Moreover, the disparity vector can be derived not only from the current reference block, but also from the blocks on the all other reference pictures included in the current reference picture lists. In addition, the disparity vector can be derived from both of the blocks predicted by disparity compensated prediction (DCP) and the blocks predicted by MCP and ARP because the derived disparity vector for ARP is stored in additional disparity vector field. Experimental results show that 0.2% bitrate reduction of synthesized views, and up to 0.5% for each dependent views in the reference software of 3D-HEVC.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115606292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Fast and smooth 3D reconstruction using multiple RGB-Depth sensors 使用多个RGB-Depth传感器进行快速平滑的3D重建
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051532
D. Alexiadis, D. Zarpalas, P. Daras
In this paper, the problem of real-time, full 3D reconstruction of foreground moving objects, an important task for Tele-Immersion applications, is addressed. More specifically, the proposed reconstruction method receives input from multiple consumer RGB-Depth cameras. A fast and efficient method to calibrate the sensors in initially described. More importantly, an efficient method to smoothly fuse the captured raw point sets is then presented, followed by a volumetric method to produce watertight and manifold meshes. Given the implementation details, the proposed method can operate at high frame rates. The experimental results, with respect to reconstruction quality and rates, verify the effectiveness of the proposed methodology.
本文研究了前景运动物体的实时、全三维重建问题,这是远程沉浸应用的一个重要任务。更具体地说,所提出的重建方法接收来自多个消费者RGB-Depth相机的输入。初步介绍了一种快速有效的传感器标定方法。更重要的是,提出了一种有效的方法来平滑融合捕获的原始点集,然后采用体积法来产生水密网格和流形网格。考虑到实现细节,所提出的方法可以在高帧率下运行。实验结果从重构质量和重构率两方面验证了所提方法的有效性。
{"title":"Fast and smooth 3D reconstruction using multiple RGB-Depth sensors","authors":"D. Alexiadis, D. Zarpalas, P. Daras","doi":"10.1109/VCIP.2014.7051532","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051532","url":null,"abstract":"In this paper, the problem of real-time, full 3D reconstruction of foreground moving objects, an important task for Tele-Immersion applications, is addressed. More specifically, the proposed reconstruction method receives input from multiple consumer RGB-Depth cameras. A fast and efficient method to calibrate the sensors in initially described. More importantly, an efficient method to smoothly fuse the captured raw point sets is then presented, followed by a volumetric method to produce watertight and manifold meshes. Given the implementation details, the proposed method can operate at high frame rates. The experimental results, with respect to reconstruction quality and rates, verify the effectiveness of the proposed methodology.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124468517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Information security display via uncrowded window 通过非拥挤窗口显示信息安全
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051604
Zhongpai Gao, Guangtao Zhai, Jiantao Zhou, Xiongkuo Min, Chunjia Hu
With the booming of visual media, people pay more and more attention to privacy protection in public environments. Most existing research on information security such as cryptography and steganography is mainly concerned about transmission and yet little has been done to prevent the information displayed on screens from reaching eyes of the bystanders. This "security of the last foot (SOLF)" problem, if left without being taken care of, will inevitably lead to the total failure of a trustable information communication system. To deal with the SOLF problem, for the application of text-reading, we proposed an eye tracking based solution using the newly revealed concept of uncrowded window from vision research. The theory of uncrowded window suggests that human vision can only effectively recognize objects inside a small window. Object features outside the window may still be detectable but the feature detection results cannot be efficiently combined properly and therefore those objects will not be recognizable. We use eye-tracker to locate fixation points of the authorized reader in real time, and only the area inside the uncrowded window displays the private information we want to protect. A number of dummy windows with fake messages are displayed around the real uncrowded window as diversions. And without the precise knowledge about the fixations of the authorized reader, the chance for bystanders to capture the private message from those surrounding area and the dummy windows is very low. Meanwhile, since the authorized reader can only read within the uncrowded window, detrimental impact of those dummy windows is almost negligible. The proposed prototype system was written in C++ with SDKs of Direct3D, Tobii Gaze SDK, CEGUI, MuPDF, OpenCV and etc. Extended demonstration of the system will be provided to show that the proposed method is an effective solution to SOLF problem of information communication and display.
随着视觉媒体的蓬勃发展,人们越来越重视公共环境中的隐私保护。大多数现有的信息安全研究,如密码学和隐写术,主要关注的是传输问题,但很少有人做过防止屏幕上显示的信息到达旁观者的眼睛。这种“最后一脚的安全(SOLF)”问题,如果不加以处理,将不可避免地导致可信赖的信息通信系统的全面失败。为了解决SOLF问题,对于文本阅读的应用,我们提出了一种基于眼动追踪的解决方案,该解决方案使用了视觉研究中新发现的非拥挤窗口的概念。非拥挤窗口理论认为人类视觉只能有效识别小窗口内的物体。窗口外的目标特征仍然可以被检测到,但特征检测结果不能有效地组合在一起,从而导致这些目标无法被识别。我们使用眼动仪实时定位授权读者的注视点,只在不拥挤的窗口内显示我们想要保护的私人信息。许多带有虚假信息的虚拟窗口被显示在真正的无人窗口周围,作为转移。如果不知道授权读者的注视点,旁观者从周围区域和虚拟窗口捕捉私人信息的机会非常低。同时,由于授权读者只能在不拥挤的窗口内阅读,这些虚拟窗口的有害影响几乎可以忽略不计。原型系统采用c++语言编写,使用Direct3D、Tobii Gaze SDK、CEGUI、MuPDF、OpenCV等SDK。系统的扩展演示表明,该方法是解决SOLF信息通信和显示问题的有效方法。
{"title":"Information security display via uncrowded window","authors":"Zhongpai Gao, Guangtao Zhai, Jiantao Zhou, Xiongkuo Min, Chunjia Hu","doi":"10.1109/VCIP.2014.7051604","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051604","url":null,"abstract":"With the booming of visual media, people pay more and more attention to privacy protection in public environments. Most existing research on information security such as cryptography and steganography is mainly concerned about transmission and yet little has been done to prevent the information displayed on screens from reaching eyes of the bystanders. This \"security of the last foot (SOLF)\" problem, if left without being taken care of, will inevitably lead to the total failure of a trustable information communication system. To deal with the SOLF problem, for the application of text-reading, we proposed an eye tracking based solution using the newly revealed concept of uncrowded window from vision research. The theory of uncrowded window suggests that human vision can only effectively recognize objects inside a small window. Object features outside the window may still be detectable but the feature detection results cannot be efficiently combined properly and therefore those objects will not be recognizable. We use eye-tracker to locate fixation points of the authorized reader in real time, and only the area inside the uncrowded window displays the private information we want to protect. A number of dummy windows with fake messages are displayed around the real uncrowded window as diversions. And without the precise knowledge about the fixations of the authorized reader, the chance for bystanders to capture the private message from those surrounding area and the dummy windows is very low. Meanwhile, since the authorized reader can only read within the uncrowded window, detrimental impact of those dummy windows is almost negligible. The proposed prototype system was written in C++ with SDKs of Direct3D, Tobii Gaze SDK, CEGUI, MuPDF, OpenCV and etc. Extended demonstration of the system will be provided to show that the proposed method is an effective solution to SOLF problem of information communication and display.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116755016","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Lossless compression of JPEG coded photo albums 无损压缩的JPEG编码相册
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051625
Hao Wu, Xiaoyan Sun, Jingyu Yang, Feng Wu
The explosion in digital photography poses a significant challenge when it comes to photo storage for both personal devices and the Internet. In this paper, we propose a novel lossless compression method to further reduce the storage size of a set of JPEG coded correlated images. In this method, we propose jointly removing the inter-image redundancy in the feature, spatial, and frequency domains. For each album, we first organize the images into a pseudo video by minimizing the global predictive cost in the feature domain. We then introduce a disparity compensation method to enhance the spatial correlation between images. Finally, the redundancy between the compensated signal and the corresponding target image is adaptively reduced in the frequency domain. Moreover, our proposed scheme is able to losslessly recover not only raw images but also JPEG files. Experimental results demonstrate the efficiency of our proposed lossless compression, which achieves more than 12% bit-saving on average compared with JPEG coded albums.
当涉及到个人设备和互联网的照片存储时,数码摄影的爆炸式增长给照片存储带来了重大挑战。在本文中,我们提出了一种新的无损压缩方法,以进一步减少一组JPEG编码的相关图像的存储大小。在该方法中,我们提出了在特征域、空间域和频域共同去除图像间冗余的方法。对于每个相册,我们首先通过最小化特征域的全局预测代价将图像组织成一个伪视频。然后引入视差补偿方法来增强图像之间的空间相关性。最后,在频域自适应降低补偿后的信号与目标图像之间的冗余。此外,我们提出的方案不仅可以无损地恢复原始图像,还可以无损地恢复JPEG文件。实验结果证明了我们提出的无损压缩方法的有效性,与JPEG编码的相册相比,平均节省了12%以上的比特。
{"title":"Lossless compression of JPEG coded photo albums","authors":"Hao Wu, Xiaoyan Sun, Jingyu Yang, Feng Wu","doi":"10.1109/VCIP.2014.7051625","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051625","url":null,"abstract":"The explosion in digital photography poses a significant challenge when it comes to photo storage for both personal devices and the Internet. In this paper, we propose a novel lossless compression method to further reduce the storage size of a set of JPEG coded correlated images. In this method, we propose jointly removing the inter-image redundancy in the feature, spatial, and frequency domains. For each album, we first organize the images into a pseudo video by minimizing the global predictive cost in the feature domain. We then introduce a disparity compensation method to enhance the spatial correlation between images. Finally, the redundancy between the compensated signal and the corresponding target image is adaptively reduced in the frequency domain. Moreover, our proposed scheme is able to losslessly recover not only raw images but also JPEG files. Experimental results demonstrate the efficiency of our proposed lossless compression, which achieves more than 12% bit-saving on average compared with JPEG coded albums.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133532929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2014 IEEE Visual Communications and Image Processing Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1