首页 > 最新文献

2013 Visual Communications and Image Processing (VCIP)最新文献

英文 中文
Compressive video sampling from a union of data-driven subspaces 从数据驱动子空间的联合压缩视频采样
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706390
Yong Li, H. Xiong, Xinwei Ye
Recently, compressive sampling (CS) is an active research field of signal processing. To further decrease the necessary measurements and get more efficient recovery of a signal x, recent approaches assume that x lives in a union of subspaces (UoS). Unlike previous approaches, this paper proposes a novel method to sample and recover an unknown signal from a union of data-driven subspaces (UoDS). Instead of a fix set of supports, this UoDS is learned from classified signal series which are uniquely formed by block matching. The basis of these data-driven subspaces is regularized after dimensionality reduction by principal component extraction. A corresponding recovery solution with provable performance guarantees is also given, which takes full advantage of block-sparsity structure and improves the recovery efficiency. In practice, the proposed scheme is fulfilled to sample and recover frames in video sequences. The experimental results demonstrate that the proposed video sampling behaves better performance in sampling and recovery than the classical CS.
压缩采样(CS)是近年来信号处理领域的一个研究热点。为了进一步减少必要的测量并更有效地恢复信号x,最近的方法假设x存在于子空间的并集(UoS)中。与以往的方法不同,本文提出了一种从数据驱动子空间并集(UoDS)中采样和恢复未知信号的新方法。该uds不是固定的一组支持,而是从通过块匹配唯一形成的分类信号序列中学习。这些数据驱动子空间的基础经过主成分提取降维后进行正则化。给出了相应的具有可证明性能保证的恢复方案,充分利用了块稀疏结构,提高了恢复效率。该方法在视频序列中实现了帧的采样和恢复。实验结果表明,所提出的视频采样方法在采样和恢复方面都优于经典的CS。
{"title":"Compressive video sampling from a union of data-driven subspaces","authors":"Yong Li, H. Xiong, Xinwei Ye","doi":"10.1109/VCIP.2013.6706390","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706390","url":null,"abstract":"Recently, compressive sampling (CS) is an active research field of signal processing. To further decrease the necessary measurements and get more efficient recovery of a signal x, recent approaches assume that x lives in a union of subspaces (UoS). Unlike previous approaches, this paper proposes a novel method to sample and recover an unknown signal from a union of data-driven subspaces (UoDS). Instead of a fix set of supports, this UoDS is learned from classified signal series which are uniquely formed by block matching. The basis of these data-driven subspaces is regularized after dimensionality reduction by principal component extraction. A corresponding recovery solution with provable performance guarantees is also given, which takes full advantage of block-sparsity structure and improves the recovery efficiency. In practice, the proposed scheme is fulfilled to sample and recover frames in video sequences. The experimental results demonstrate that the proposed video sampling behaves better performance in sampling and recovery than the classical CS.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"191 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123533287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised color classifier training for soccer player detection 足球运动员检测的无监督颜色分类器训练
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706424
S. Gerke, Shiwang Singh, A. Linnemann, P. Ndjiki-Nya
Player detection in sports video is a challenging task: In contrast to typical surveillance applications, a pan-tilt-zoom camera model is used. Therefore, simple background learning approaches cannot be used. Furthermore, camera motion causes severe motion blur, making gradient based approaches less robust than in settings where the camera is static. The contribution of this paper is a sequence adaptive approach that utilizes color information in an unsupervised manner to improve detection accuracy. Therefore, different color features, namely color histograms, color spatiograms and a color and edge directivity descriptor are evaluated. It is shown that the proposed color adaptive approach improves detection accuracy. In terms of maximum F1 score, an improvement from 0.79 to 0.81 is reached using block-wise HSV histograms. The average number of false positives per image (FPPI) at two fixed recall levels decreased by approximately 23%.
体育视频中的球员检测是一项具有挑战性的任务:与典型的监控应用相比,使用了一种泛倾斜变焦相机模型。因此,不能使用简单的背景学习方法。此外,相机运动导致严重的运动模糊,使基于梯度的方法不如相机静态设置的鲁棒性。本文的贡献是一种序列自适应方法,该方法以无监督的方式利用颜色信息来提高检测精度。因此,评估不同的颜色特征,即颜色直方图,颜色空间图以及颜色和边缘指向性描述符。结果表明,所提出的颜色自适应方法提高了检测精度。在最大F1分数方面,使用分块HSV直方图可以从0.79提高到0.81。在两个固定召回水平下,每张图像的平均误报数(FPPI)下降了约23%。
{"title":"Unsupervised color classifier training for soccer player detection","authors":"S. Gerke, Shiwang Singh, A. Linnemann, P. Ndjiki-Nya","doi":"10.1109/VCIP.2013.6706424","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706424","url":null,"abstract":"Player detection in sports video is a challenging task: In contrast to typical surveillance applications, a pan-tilt-zoom camera model is used. Therefore, simple background learning approaches cannot be used. Furthermore, camera motion causes severe motion blur, making gradient based approaches less robust than in settings where the camera is static. The contribution of this paper is a sequence adaptive approach that utilizes color information in an unsupervised manner to improve detection accuracy. Therefore, different color features, namely color histograms, color spatiograms and a color and edge directivity descriptor are evaluated. It is shown that the proposed color adaptive approach improves detection accuracy. In terms of maximum F1 score, an improvement from 0.79 to 0.81 is reached using block-wise HSV histograms. The average number of false positives per image (FPPI) at two fixed recall levels decreased by approximately 23%.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124929135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Multi-model prediction for image set compression 图像集压缩的多模型预测
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706334
Zhongbo Shi, Xiaoyan Sun, Feng Wu
The key task in image set compression is how to efficiently remove set redundancy among images and within a single image. In this paper, we propose the first multi-model prediction (MoP) method for image set compression to significantly reduce inter image redundancy. Unlike the previous prediction methods, our MoP enhances the correlation between images using feature-based geometric multi-model fitting. Based on estimated geometric models, multiple deformed prediction images are generated to reduce geometric distortions in different image regions. The block-based adaptive motion compensation is then adopted to further eliminate local variances. Experimental results demonstrate the advantage of our approach, especially for images with complicated scenes and geometric relationships.
图像集压缩的关键问题是如何有效地去除图像之间和单个图像内的集冗余。在本文中,我们提出了第一种用于图像集压缩的多模型预测(MoP)方法,以显着降低图像间冗余。与之前的预测方法不同,我们的MoP使用基于特征的几何多模型拟合来增强图像之间的相关性。基于估计的几何模型,生成多个变形预测图像,以减少不同图像区域的几何畸变。然后采用基于分块的自适应运动补偿进一步消除局部方差。实验结果证明了该方法的优越性,特别是对于具有复杂场景和几何关系的图像。
{"title":"Multi-model prediction for image set compression","authors":"Zhongbo Shi, Xiaoyan Sun, Feng Wu","doi":"10.1109/VCIP.2013.6706334","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706334","url":null,"abstract":"The key task in image set compression is how to efficiently remove set redundancy among images and within a single image. In this paper, we propose the first multi-model prediction (MoP) method for image set compression to significantly reduce inter image redundancy. Unlike the previous prediction methods, our MoP enhances the correlation between images using feature-based geometric multi-model fitting. Based on estimated geometric models, multiple deformed prediction images are generated to reduce geometric distortions in different image regions. The block-based adaptive motion compensation is then adopted to further eliminate local variances. Experimental results demonstrate the advantage of our approach, especially for images with complicated scenes and geometric relationships.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127850529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
An adaptive texture-depth rate allocation estimation technique for low latency multi-view video plus depth transmission 一种低延迟多视点视频加深度传输的自适应纹理深度率分配估计技术
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706332
M. Cordina, C. J. Debono
This paper presents an adaptive texture-depth target bit rate allocation estimation technique for low latency multi-view video plus depth transmission using a multi-regression model. The proposed technique employs the prediction mode distribution of the macroblocks at the discontinuity regions of the depth map video to estimate the optimal texture-depth target bit rate allocation considering the total available bit rate. This technique was tested using various standard test sequences and has shown efficacy as the model is able to estimate, in real-time, the optimal texture-depth rate allocation with an absolute mean estimation error of 2.5% and a standard deviation of 2.2%. Moreover, it allows the texture-depth rate allocation to be adapted to the video sequence with good tracking performance, allowing the correct handling of scene changes.
本文提出了一种基于多元回归模型的低延迟多视点视频加深度传输的自适应纹理深度目标比特率分配估计技术。该技术利用深度图视频不连续区域宏块的预测模式分布,在考虑总可用比特率的情况下估计最优纹理-深度目标比特率分配。该方法在不同的标准测试序列中进行了测试,结果表明,该模型能够实时估计最佳纹理-深度率分配,绝对平均估计误差为2.5%,标准差为2.2%。此外,它允许纹理深度率分配适应视频序列具有良好的跟踪性能,允许正确处理场景的变化。
{"title":"An adaptive texture-depth rate allocation estimation technique for low latency multi-view video plus depth transmission","authors":"M. Cordina, C. J. Debono","doi":"10.1109/VCIP.2013.6706332","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706332","url":null,"abstract":"This paper presents an adaptive texture-depth target bit rate allocation estimation technique for low latency multi-view video plus depth transmission using a multi-regression model. The proposed technique employs the prediction mode distribution of the macroblocks at the discontinuity regions of the depth map video to estimate the optimal texture-depth target bit rate allocation considering the total available bit rate. This technique was tested using various standard test sequences and has shown efficacy as the model is able to estimate, in real-time, the optimal texture-depth rate allocation with an absolute mean estimation error of 2.5% and a standard deviation of 2.2%. Moreover, it allows the texture-depth rate allocation to be adapted to the video sequence with good tracking performance, allowing the correct handling of scene changes.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116432004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Joint video/depth/FEC rate allocation with considering 3D visual saliency for scalable 3D video streaming 联合视频/深度/FEC速率分配,考虑可扩展3D视频流的3D视觉显着性
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706339
Yanwei Liu, Jinxia Liu, S. Ci, Yun Ye
For robust video plus depth based 3D video streaming, video, depth and packet-level forward error correction (FEC) can provide many rate combinations with various 3D visual qualities to adapt to the dynamic channel conditions. Video/depth/FEC rate allocation under the channel bandwidth constraint is an important optimization problem for robust 3D video streaming. This paper proposes a joint video/depth/FEC rate allocation method by maximizing the receiver's 3D visual quality. Through predicting the perceptual 3D visual qualities of the different video/depth/FEC rate combinations, the optimal GOP-level video/depth/FEC rate combination can be found. Further, the selected FEC rates are unequally assigned to different levels of 3D saliency regions within each video/depth frame. The effectiveness of the proposed 3D saliency based joint video/depth/FEC rate allocation method for scalable 3D video streaming is validated by extensive experimental results.
对于鲁棒的基于视频和深度的3D视频流,视频、深度和包级前向纠错(FEC)可以提供多种具有不同3D视觉质量的速率组合,以适应动态信道条件。信道带宽约束下的视频/深度/FEC速率分配是鲁棒3D视频流的一个重要优化问题。本文提出了一种最大化接收机三维视觉质量的视频/深度/FEC联合速率分配方法。通过预测不同视频/深度/FEC速率组合的感知3D视觉质量,可以找到最佳gop级视频/深度/FEC速率组合。此外,所选择的FEC速率不均匀地分配到每个视频/深度帧内不同级别的3D显著性区域。大量的实验结果验证了所提出的基于3D显著性的视频/深度/FEC速率联合分配方法在可扩展3D视频流中的有效性。
{"title":"Joint video/depth/FEC rate allocation with considering 3D visual saliency for scalable 3D video streaming","authors":"Yanwei Liu, Jinxia Liu, S. Ci, Yun Ye","doi":"10.1109/VCIP.2013.6706339","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706339","url":null,"abstract":"For robust video plus depth based 3D video streaming, video, depth and packet-level forward error correction (FEC) can provide many rate combinations with various 3D visual qualities to adapt to the dynamic channel conditions. Video/depth/FEC rate allocation under the channel bandwidth constraint is an important optimization problem for robust 3D video streaming. This paper proposes a joint video/depth/FEC rate allocation method by maximizing the receiver's 3D visual quality. Through predicting the perceptual 3D visual qualities of the different video/depth/FEC rate combinations, the optimal GOP-level video/depth/FEC rate combination can be found. Further, the selected FEC rates are unequally assigned to different levels of 3D saliency regions within each video/depth frame. The effectiveness of the proposed 3D saliency based joint video/depth/FEC rate allocation method for scalable 3D video streaming is validated by extensive experimental results.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125598265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A bank of fast matched filters by decomposing the filter kernel 通过分解滤波器核得到一组快速匹配的滤波器
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706434
Mihails Pudzs, Rihards Fuksis, M. Greitans, Teodors Eglitis
In this paper we introduce a bank of fast matched filters that are designed to extract gradients, edges, lines and various line crossings. Our work is based on previously introduced filtering approaches like conventional Matched Filtering (MF), Complex Matched Filtering (CMF) and Generalized Complex Matched Filtering (GCMF), and is aimed to speed up the image processing. Filter kernel decomposition method is demonstrated for the latter mentioned (GCMF) but can be similarly applied to any other filters (like MF, CMF, Gabor filters, spiculation filters, steerable MF, etc.) as well. By introducing the mask kernel approximation, we show how to substitute the GCMF with several more computationally efficient filters, which reduce the overall computation complexity by over hundred of times. Acquired Fast GCMF retains all of the functionality of GCMF (extracts the desired objects and obtains their angular orientation), losing in accuracy only about +26 dB in terms of PSNR.
本文介绍了一组快速匹配滤波器,用于提取梯度、边缘、直线和各种直线交叉。我们的工作基于先前介绍的滤波方法,如传统匹配滤波(MF),复杂匹配滤波(CMF)和广义复杂匹配滤波(GCMF),旨在加快图像处理速度。对于后者(GCMF)演示了滤波器核分解方法,但也可以类似地应用于任何其他滤波器(如MF, CMF, Gabor滤波器,spiculation滤波器,可操纵MF等)。通过引入掩模核近似,我们展示了如何用几个计算效率更高的滤波器代替GCMF,从而将总体计算复杂度降低了数百倍以上。获得的快速GCMF保留了GCMF的所有功能(提取所需物体并获得它们的角方向),在PSNR方面精度仅损失约26 dB。
{"title":"A bank of fast matched filters by decomposing the filter kernel","authors":"Mihails Pudzs, Rihards Fuksis, M. Greitans, Teodors Eglitis","doi":"10.1109/VCIP.2013.6706434","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706434","url":null,"abstract":"In this paper we introduce a bank of fast matched filters that are designed to extract gradients, edges, lines and various line crossings. Our work is based on previously introduced filtering approaches like conventional Matched Filtering (MF), Complex Matched Filtering (CMF) and Generalized Complex Matched Filtering (GCMF), and is aimed to speed up the image processing. Filter kernel decomposition method is demonstrated for the latter mentioned (GCMF) but can be similarly applied to any other filters (like MF, CMF, Gabor filters, spiculation filters, steerable MF, etc.) as well. By introducing the mask kernel approximation, we show how to substitute the GCMF with several more computationally efficient filters, which reduce the overall computation complexity by over hundred of times. Acquired Fast GCMF retains all of the functionality of GCMF (extracts the desired objects and obtains their angular orientation), losing in accuracy only about +26 dB in terms of PSNR.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133947618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Perceptual grouping via untangling Gestalt principles 通过解缠格式塔原则的知觉分组
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706384
Yonggang Qi, Jun Guo, Yi Li, Honggang Zhang, T. Xiang, Yi-Zhe Song, Z. Tan
Gestalt principles, a set of conjoining rules derived from human visual studies, have been known to play an important role in computer vision. Many applications such as image segmentation, contour grouping and scene understanding often rely on such rules to work. However, the problem of Gestalt confliction, i.e., the relative importance of each rule compared with another, remains unsolved. In this paper, we investigate the problem of perceptual grouping by quantifying the confliction among three commonly used rules: similarity, continuity and proximity. More specifically, we propose to quantify the importance of Gestalt rules by solving a learning to rank problem, and formulate a multi-label graph-cuts algorithm to group image primitives while taking into account the learned Gestalt confliction. Our experiment results confirm the existence of Gestalt confliction in perceptual grouping and demonstrate an improved performance when such a confliction is accounted for via the proposed grouping algorithm. Finally, a novel cross domain image classification method is proposed by exploiting perceptual grouping as representation.
格式塔原则是一套源自人类视觉研究的连接规则,在计算机视觉中发挥着重要作用。许多应用,如图像分割,轮廓分组和场景理解往往依赖于这些规则的工作。然而,格式塔冲突的问题,即每个规则相对于另一个规则的相对重要性,仍然没有得到解决。本文通过量化三种常用规则:相似性、连续性和接近性之间的冲突来研究感知分组问题。更具体地说,我们建议通过解决一个学习排序问题来量化格式塔规则的重要性,并制定一个多标签图切割算法来对图像原语进行分组,同时考虑到学习到的格式塔冲突。我们的实验结果证实了感知分组中格式塔冲突的存在,并证明了通过所提出的分组算法考虑这种冲突后的性能有所提高。最后,提出了一种利用感知分组作为表示的跨域图像分类方法。
{"title":"Perceptual grouping via untangling Gestalt principles","authors":"Yonggang Qi, Jun Guo, Yi Li, Honggang Zhang, T. Xiang, Yi-Zhe Song, Z. Tan","doi":"10.1109/VCIP.2013.6706384","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706384","url":null,"abstract":"Gestalt principles, a set of conjoining rules derived from human visual studies, have been known to play an important role in computer vision. Many applications such as image segmentation, contour grouping and scene understanding often rely on such rules to work. However, the problem of Gestalt confliction, i.e., the relative importance of each rule compared with another, remains unsolved. In this paper, we investigate the problem of perceptual grouping by quantifying the confliction among three commonly used rules: similarity, continuity and proximity. More specifically, we propose to quantify the importance of Gestalt rules by solving a learning to rank problem, and formulate a multi-label graph-cuts algorithm to group image primitives while taking into account the learned Gestalt confliction. Our experiment results confirm the existence of Gestalt confliction in perceptual grouping and demonstrate an improved performance when such a confliction is accounted for via the proposed grouping algorithm. Finally, a novel cross domain image classification method is proposed by exploiting perceptual grouping as representation.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"26 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133389662","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Lossless predictive coding with Bayesian treatment 采用贝叶斯处理的无损预测编码
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706328
Jing Liu, Xiaokang Yang, Guangtao Zhai, Li Chen, Xianghui Sun, Wanhong Chen, Ying Zuo
Natural image statistics have been widely exploited for lossless predictive coding and other applications. However, traditional adaptive techniques always focus on the local consistency of training set regardless of what the predicted target looks like. We investigate the problem of introducing the model evidence of predicted target since self-similarity inherent in natural images gives some kind of prior information for the distribution of predicted result. The proposed Bayesian model integrated with both training evidence and target evidence takes full advantages of local structure as well as self-similarity. Experimental results demonstrate that the proposed context model achieves best results compared with the state-of-the-art lossless predictors.
自然图像统计已广泛用于无损预测编码和其他应用。然而,传统的自适应技术总是关注训练集的局部一致性,而不管预测目标是什么样子。由于自然图像固有的自相似性为预测结果的分布提供了某种先验信息,我们研究了引入预测目标模型证据的问题。所提出的训练证据和目标证据相结合的贝叶斯模型充分利用了局部结构和自相似性的优点。实验结果表明,与目前最先进的无损预测器相比,本文提出的上下文模型取得了最好的效果。
{"title":"Lossless predictive coding with Bayesian treatment","authors":"Jing Liu, Xiaokang Yang, Guangtao Zhai, Li Chen, Xianghui Sun, Wanhong Chen, Ying Zuo","doi":"10.1109/VCIP.2013.6706328","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706328","url":null,"abstract":"Natural image statistics have been widely exploited for lossless predictive coding and other applications. However, traditional adaptive techniques always focus on the local consistency of training set regardless of what the predicted target looks like. We investigate the problem of introducing the model evidence of predicted target since self-similarity inherent in natural images gives some kind of prior information for the distribution of predicted result. The proposed Bayesian model integrated with both training evidence and target evidence takes full advantages of local structure as well as self-similarity. Experimental results demonstrate that the proposed context model achieves best results compared with the state-of-the-art lossless predictors.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133402841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Accurate 3D reconstruction of dynamic scenes with Fourier transform assisted phase shifting 傅里叶变换辅助相移的动态场景精确三维重建
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706399
Pengyu Cong, Yueyi Zhang, Zhiwei Xiong, Shenghui Zhao, Feng Wu
Phase shifting is a widely used method for accurate and dense 3D reconstruction. However, at least three images of the same scene are required for each reconstruction, so measurement errors are inevitable in dynamic scenes, even with high-speed hardware. In this paper, we propose a Fourier transform assisted phase shifting method to overcome the motion vulnerability in phase shifting. A new model with motion-related phase shifts is formulated, and the coarse phase measurements obtained by Fourier transform profilemetry are used to estimate the unknown phase shifts. The phase errors caused by motion are greatly reduced in this way. Experimental results show that the proposed method can obtain accurate and dense 3D reconstruction of dynamic scenes, with regard to different kinds of motion.
相移是一种广泛使用的精确、密集三维重建方法。然而,每次重建至少需要三张相同场景的图像,因此在动态场景中,即使使用高速硬件,测量误差也是不可避免的。本文提出了一种傅里叶变换辅助相移方法来克服相移中的运动脆弱性。建立了一个具有运动相关相移的新模型,并利用傅里叶变换轮廓法得到的粗相测量值来估计未知相移。这种方法大大减小了运动引起的相位误差。实验结果表明,该方法可以对不同运动类型的动态场景进行精确、密集的三维重建。
{"title":"Accurate 3D reconstruction of dynamic scenes with Fourier transform assisted phase shifting","authors":"Pengyu Cong, Yueyi Zhang, Zhiwei Xiong, Shenghui Zhao, Feng Wu","doi":"10.1109/VCIP.2013.6706399","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706399","url":null,"abstract":"Phase shifting is a widely used method for accurate and dense 3D reconstruction. However, at least three images of the same scene are required for each reconstruction, so measurement errors are inevitable in dynamic scenes, even with high-speed hardware. In this paper, we propose a Fourier transform assisted phase shifting method to overcome the motion vulnerability in phase shifting. A new model with motion-related phase shifts is formulated, and the coarse phase measurements obtained by Fourier transform profilemetry are used to estimate the unknown phase shifts. The phase errors caused by motion are greatly reduced in this way. Experimental results show that the proposed method can obtain accurate and dense 3D reconstruction of dynamic scenes, with regard to different kinds of motion.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129358096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Entropy of primitive: A top-down methodology for evaluating the perceptual visual information 原语熵:一种自顶向下评估感性视觉信息的方法
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706358
Xianguo Zhang, Shiqi Wang, Siwei Ma, Shaohui Liu, Wen Gao
In this paper, we aim at evaluating the perceptual visual information based on a novel top-down methodology: entropy of primitive (EoP). The EoP is determined by the distribution of the atoms in describing an image, and is demonstrated to exhibit closely correlation with the perceptual image quality. Based on the visual information evaluation, we further demonstrate that the EoP is effective in predicting the perceptual lossless of natural images. Inspired by this observation, in order to distinguish whether the loss of input signal is visual noticeable to human visual system (HVS), we introduce the EoP based perceptual lossless profile (PLP). Extensive experiments verify that, the proposed EoP based perceptual lossless profile can efficiently measure the minimum noticeable visual information distortion and achieve better performance compared to the-state-of-the-art just-noticeable difference (JND) profile.
本文基于一种新颖的自顶向下的方法:原始熵(EoP)来评估感知视觉信息。EoP由描述图像的原子分布决定,并被证明与感知图像质量密切相关。在视觉信息评价的基础上,我们进一步证明了EoP在预测自然图像的感知无损方面是有效的。受此启发,为了区分输入信号的损失对人类视觉系统(HVS)是否可见,我们引入了基于EoP的感知无损轮廓(PLP)。大量实验证明,本文提出的基于EoP的感知无损轮廓可以有效地测量最小可注意视觉信息失真,并且与最新的刚可注意差分(JND)轮廓相比,具有更好的性能。
{"title":"Entropy of primitive: A top-down methodology for evaluating the perceptual visual information","authors":"Xianguo Zhang, Shiqi Wang, Siwei Ma, Shaohui Liu, Wen Gao","doi":"10.1109/VCIP.2013.6706358","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706358","url":null,"abstract":"In this paper, we aim at evaluating the perceptual visual information based on a novel top-down methodology: entropy of primitive (EoP). The EoP is determined by the distribution of the atoms in describing an image, and is demonstrated to exhibit closely correlation with the perceptual image quality. Based on the visual information evaluation, we further demonstrate that the EoP is effective in predicting the perceptual lossless of natural images. Inspired by this observation, in order to distinguish whether the loss of input signal is visual noticeable to human visual system (HVS), we introduce the EoP based perceptual lossless profile (PLP). Extensive experiments verify that, the proposed EoP based perceptual lossless profile can efficiently measure the minimum noticeable visual information distortion and achieve better performance compared to the-state-of-the-art just-noticeable difference (JND) profile.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"29 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132407395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
期刊
2013 Visual Communications and Image Processing (VCIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1