首页 > 最新文献

2014 IEEE Visual Communications and Image Processing Conference最新文献

英文 中文
Tag-based social image search with hyperedges correlation 基于标签的超边缘关联社交图像搜索
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051573
Leiquan Wang, Zhicheng Zhao, Fei Su
In social image search, most existing hypergraph methods use the visual and textual features in isolation by treating each feature term as a hyperedge. Nevertheless, they neglect the correlations of visual and textual hyperedges, which are more robust to represent the high-order relationship among vertices. In this paper, we propose a hypergraph with correlated hyperedges (CHH), which introduces high-order relationship of hyperedges into hypergraph learning. Based on CHH, a pairwise visual-textual correlation hypergraph (VTCH) model is used for tag-based social image search. To overcome the large number of newly generated hybrid hyperedges, a bagging-based method is adopted to balance the accuracy and speed. Finally, adaptive hyperedges learning method is used to obtain the relevance score for social image search. The experiments conducted on MIR Flickr show the effectiveness of our proposed method.
在社交图像搜索中,大多数现有的超图方法通过将每个特征项视为超边缘来孤立地使用视觉和文本特征。然而,它们忽略了视觉和文本超边缘的相关性,这对于表示顶点之间的高阶关系来说更健壮。本文提出了一种具有相关超边的超图(CHH),将超边的高阶关系引入到超图学习中。在此基础上,提出了一种基于标签的视觉文本相关超图(VTCH)模型。为了克服新生成的混合超边数量多的问题,采用了一种基于装袋的方法来平衡精度和速度。最后,采用自适应超边缘学习方法获得社交图像搜索的相关分数。在MIR Flickr上进行的实验表明了该方法的有效性。
{"title":"Tag-based social image search with hyperedges correlation","authors":"Leiquan Wang, Zhicheng Zhao, Fei Su","doi":"10.1109/VCIP.2014.7051573","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051573","url":null,"abstract":"In social image search, most existing hypergraph methods use the visual and textual features in isolation by treating each feature term as a hyperedge. Nevertheless, they neglect the correlations of visual and textual hyperedges, which are more robust to represent the high-order relationship among vertices. In this paper, we propose a hypergraph with correlated hyperedges (CHH), which introduces high-order relationship of hyperedges into hypergraph learning. Based on CHH, a pairwise visual-textual correlation hypergraph (VTCH) model is used for tag-based social image search. To overcome the large number of newly generated hybrid hyperedges, a bagging-based method is adopted to balance the accuracy and speed. Finally, adaptive hyperedges learning method is used to obtain the relevance score for social image search. The experiments conducted on MIR Flickr show the effectiveness of our proposed method.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"363 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121723298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Region-of-unpredictable determination for accelerated full-frame feature generation in video sequences 视频序列中加速全帧特征生成的不可预测区域确定
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051599
Jia-Lin Chen, Chun-Chen Kuo, Liang-Gee Chen
In this paper, we propose a novel concept of region-of-unpredictable (ROU) to accelerate full-frame feature generation in video sequences. Due to the high correlation between successive frames, there are only few regions in which the features could not be estimated accurately from the previous frame called region-of-unpredictable (ROU). We develop a scheme combining partial feature extraction in ROU with feature prediction from the previous frame. The full-frame features of the current frame can then be obtained to minimize information loss. Experimental results show that the ROU determination algorithm supports 95.71% detection rate. The full-frame feature generation scheme using ROU determination saves 79.38% computational time compared with the full-frame feature extraction.
在本文中,我们提出了一种新的不可预测区域(ROU)概念来加速视频序列中全帧特征的生成。由于连续帧之间的高度相关性,只有少数区域的特征不能从前一帧中准确估计出来,称为不可预测区域(ROU)。我们开发了一种将ROU中的部分特征提取与前一帧的特征预测相结合的方案。然后可以获得当前帧的全帧特征,以最小化信息损失。实验结果表明,该ROU确定算法的检测率为95.71%。采用ROU确定的全帧特征生成方案与全帧特征提取方案相比,计算时间节省79.38%。
{"title":"Region-of-unpredictable determination for accelerated full-frame feature generation in video sequences","authors":"Jia-Lin Chen, Chun-Chen Kuo, Liang-Gee Chen","doi":"10.1109/VCIP.2014.7051599","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051599","url":null,"abstract":"In this paper, we propose a novel concept of region-of-unpredictable (ROU) to accelerate full-frame feature generation in video sequences. Due to the high correlation between successive frames, there are only few regions in which the features could not be estimated accurately from the previous frame called region-of-unpredictable (ROU). We develop a scheme combining partial feature extraction in ROU with feature prediction from the previous frame. The full-frame features of the current frame can then be obtained to minimize information loss. Experimental results show that the ROU determination algorithm supports 95.71% detection rate. The full-frame feature generation scheme using ROU determination saves 79.38% computational time compared with the full-frame feature extraction.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123302418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast intra partition algorithm for HEVC screen content coding HEVC屏幕内容编码的快速帧内分割算法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051588
Mengmeng Zhang, Yuhui Guo, H. Bai
Since the publication of the High Efficiency Video Coding standard as the newest video coding standard, several extensions have been made. Among these, the use of the screen content coding in many fields is one of the important extensions. In terms of coding tree unit (CTU) partitioning, rate distortion optimization is still used in screen content coding. The complexity of the process has resulted in problems in relation to real-time application. Thus, this paper proposes a fast-deciding CTU partition mode algorithm based on entropy and coding bits. Experimental results show that the proposed algorithm can save 32% of encoding time on average compared with the default algorithm in HM-12.1+RExt-5.1 with only 0.8% bit rate increment in coding performance.
高效视频编码标准作为最新的视频编码标准发布以来,已经进行了多次扩展。其中,屏幕内容编码在许多领域的应用是重要的扩展之一。在编码树单元(CTU)划分方面,屏幕内容编码仍然采用率失真优化。该过程的复杂性导致了与实时应用相关的问题。为此,本文提出了一种基于熵和编码位的快速确定CTU划分模式的算法。实验结果表明,与HM-12.1+ ext -5.1的默认算法相比,该算法平均可节省32%的编码时间,编码性能仅提高0.8%。
{"title":"Fast intra partition algorithm for HEVC screen content coding","authors":"Mengmeng Zhang, Yuhui Guo, H. Bai","doi":"10.1109/VCIP.2014.7051588","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051588","url":null,"abstract":"Since the publication of the High Efficiency Video Coding standard as the newest video coding standard, several extensions have been made. Among these, the use of the screen content coding in many fields is one of the important extensions. In terms of coding tree unit (CTU) partitioning, rate distortion optimization is still used in screen content coding. The complexity of the process has resulted in problems in relation to real-time application. Thus, this paper proposes a fast-deciding CTU partition mode algorithm based on entropy and coding bits. Experimental results show that the proposed algorithm can save 32% of encoding time on average compared with the default algorithm in HM-12.1+RExt-5.1 with only 0.8% bit rate increment in coding performance.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116917506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Two-dimensional histogram expansion of wavelet coefficient for reversible data hiding 可逆数据隐藏的小波系数二维直方图展开
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051553
Kazuki Yamato, Kazuma Shinoda, Madoka Hasegawa, Shigeo Kato
In this paper, we propose a reversible data hiding (RDH) method based on a two-dimensional wavelet coefficient histogram (2D WCH) in the wavelet domain. First, a cover image is decomposed into wavelet subbands using the invertible integer-to-integer wavelet transform (121-WT). Then, the 2D WCH is generated by counting the occurrence frequency of the wavelet coefficient pairs which denote two wavelet coefficients located in the same position in the selected two subbands where the secret message is embedded. By using the 2D WCH, the correlation between the selected subbands is more effectively utilized than the traditional ID histogram. The proposed method embed the secret message reversibly in the cover image by expanding the 2D WCH. In order to embed the secret message as efficient as possible, the expansion rule for 2D WCH is proposed. Moreover, the coefficient pair selection (CPS), which the coefficients embedding the data are selected in order to modify only the selected coefficients, is implemented before generating the 2D WCH. In the experiment, the proposed method is compared with the conventional RDH methods in terms of the capacity-distortion curve.
本文提出了一种基于小波域二维小波系数直方图(2D WCH)的可逆数据隐藏方法。首先,利用可逆整数到整数小波变换(121-WT)将封面图像分解成小波子带;然后,通过计算小波系数对的出现频率来生成二维WCH,这些小波系数对表示在选定的嵌入秘密信息的两个子带中位于相同位置的两个小波系数。与传统的ID直方图相比,使用二维WCH可以更有效地利用所选子带之间的相关性。该方法通过扩展二维WCH,将秘密信息可逆嵌入到封面图像中。为了尽可能高效地嵌入秘密信息,提出了二维WCH的展开规则。在生成二维WCH之前,实现了系数对选择(CPS),即选择嵌入数据的系数,只修改被选中的系数。在实验中,将该方法与传统的RDH方法在容量畸变曲线上进行了比较。
{"title":"Two-dimensional histogram expansion of wavelet coefficient for reversible data hiding","authors":"Kazuki Yamato, Kazuma Shinoda, Madoka Hasegawa, Shigeo Kato","doi":"10.1109/VCIP.2014.7051553","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051553","url":null,"abstract":"In this paper, we propose a reversible data hiding (RDH) method based on a two-dimensional wavelet coefficient histogram (2D WCH) in the wavelet domain. First, a cover image is decomposed into wavelet subbands using the invertible integer-to-integer wavelet transform (121-WT). Then, the 2D WCH is generated by counting the occurrence frequency of the wavelet coefficient pairs which denote two wavelet coefficients located in the same position in the selected two subbands where the secret message is embedded. By using the 2D WCH, the correlation between the selected subbands is more effectively utilized than the traditional ID histogram. The proposed method embed the secret message reversibly in the cover image by expanding the 2D WCH. In order to embed the secret message as efficient as possible, the expansion rule for 2D WCH is proposed. Moreover, the coefficient pair selection (CPS), which the coefficients embedding the data are selected in order to modify only the selected coefficients, is implemented before generating the 2D WCH. In the experiment, the proposed method is compared with the conventional RDH methods in terms of the capacity-distortion curve.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121153626","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Improving a vision indoor localization system by a saliency-guided detection 利用显著性引导检测改进视觉室内定位系统
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051526
Wael Elloumi, Kamel Guissous, A. Chetouani, S. Treuillet
In this paper, we propose to use visual saliency to improve an indoor localization system based on image matching. A learning step permits to determinate the reference trajectory by selecting some key frames along the path. During the localization step, the current image is then compared to the obtained key frames in order to estimate the user's position. This comparison is realized by extracting primitive information through a saliency method, which aims to improve our localization system by focusing our attention on the more singular regions to match. Another advantage of the saliency-guided detection is to save computation time. The proposed framework has been developed and tested on a Smartphone. The obtained results show the interest of the use of saliency models by comparing the numbers of features and good matches in video sequence.
本文提出利用视觉显著性来改进基于图像匹配的室内定位系统。学习步骤允许通过沿着路径选择一些关键帧来确定参考轨迹。在定位步骤中,将当前图像与获得的关键帧进行比较,以估计用户的位置。这种比较是通过显著性方法提取原始信息来实现的,该方法旨在通过将注意力集中在更奇异的区域上来改进我们的定位系统。显著性引导检测的另一个优点是节省计算时间。提出的框架已经在智能手机上开发和测试。通过比较视频序列中特征的数量和良好的匹配,得到了显著性模型的应用效果。
{"title":"Improving a vision indoor localization system by a saliency-guided detection","authors":"Wael Elloumi, Kamel Guissous, A. Chetouani, S. Treuillet","doi":"10.1109/VCIP.2014.7051526","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051526","url":null,"abstract":"In this paper, we propose to use visual saliency to improve an indoor localization system based on image matching. A learning step permits to determinate the reference trajectory by selecting some key frames along the path. During the localization step, the current image is then compared to the obtained key frames in order to estimate the user's position. This comparison is realized by extracting primitive information through a saliency method, which aims to improve our localization system by focusing our attention on the more singular regions to match. Another advantage of the saliency-guided detection is to save computation time. The proposed framework has been developed and tested on a Smartphone. The obtained results show the interest of the use of saliency models by comparing the numbers of features and good matches in video sequence.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121107600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Robust 3D LUT estimation method for SHVC color gamut scalability SHVC色域可扩展性的鲁棒3D LUT估计方法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051510
Yuwen He, Yan Ye, Jie Dong
Color gamut scalability (CGS) in scalable extensions of High Efficiency Video Coding (SHVC) supports scalable coding with multiple layers in different color spaces. Base layer conveying HDTV video in BT.709 color space and enhancement layer conveying UHDTV video in BT.2020 color space is identified as a practical use case for CGS. Efficient CGS coding can be achieved using a 3D Look-up Table (LUT) based color conversion process. This paper proposes a robust 3D LUT parameter estimation method that estimates the 3D LUT parameters globally using the Least Square method. Problems of matrix sparsity and uneven sample distribution are carefully handled to improve the stability and accuracy of the estimation process. Simulation results confirm that the proposed 3D LUT estimation method can significantly improve coding performance compared with other gamut conversion methods.
高效视频编码(High Efficiency Video Coding, SHVC)的可扩展扩展中的色域可扩展性(CGS)支持在不同颜色空间中进行多层可扩展编码。确定了在BT.709色彩空间中传输HDTV视频的基层和在BT.2020色彩空间中传输UHDTV视频的增强层作为CGS的实际用例。使用基于3D查找表(LUT)的颜色转换过程可以实现高效的CGS编码。提出了一种鲁棒的三维LUT参数估计方法,利用最小二乘法对三维LUT参数进行全局估计。为了提高估计过程的稳定性和准确性,对矩阵稀疏性和样本分布不均匀等问题进行了细致的处理。仿真结果表明,与其他色域转换方法相比,所提出的三维LUT估计方法可以显著提高编码性能。
{"title":"Robust 3D LUT estimation method for SHVC color gamut scalability","authors":"Yuwen He, Yan Ye, Jie Dong","doi":"10.1109/VCIP.2014.7051510","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051510","url":null,"abstract":"Color gamut scalability (CGS) in scalable extensions of High Efficiency Video Coding (SHVC) supports scalable coding with multiple layers in different color spaces. Base layer conveying HDTV video in BT.709 color space and enhancement layer conveying UHDTV video in BT.2020 color space is identified as a practical use case for CGS. Efficient CGS coding can be achieved using a 3D Look-up Table (LUT) based color conversion process. This paper proposes a robust 3D LUT parameter estimation method that estimates the 3D LUT parameters globally using the Least Square method. Problems of matrix sparsity and uneven sample distribution are carefully handled to improve the stability and accuracy of the estimation process. Simulation results confirm that the proposed 3D LUT estimation method can significantly improve coding performance compared with other gamut conversion methods.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115380796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Blind image quality assessment based on a new feature of nature scene statistics 基于自然场景统计新特征的盲图像质量评价
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051498
Li Song, Chen Chen, Yi Xu, Genjian Xue, Yi Zhou
A recently proposed model, known as blind/referenceless image spatial quality evaluator (BRISQUE), achieves the state-of-the-art performance in context of blind image quality assessment (IQA). This model used the predefined generalized Gaussian distribution (GGD) to describe the regularity of natural scene statistics, introducing fitting errors due to variations of image contents. In this paper, a more generalized model is proposed to better characterize the regularity of extensive image contents, which is learned from the concatenated histograms of mean subtracted contrast normalized (MSCN) coefficients and pairwise products of MSCN coefficients of neighbouring pixels. The new feature based on MSCN shows its capability of preserving intrinsic distribution of image statistics. Consequently support vector machine regression (SVR) can map it to more accurate image quality scores. Experimental results show that the proposed approach achieves a slight gain from BRISQUE, which indicates the crafted GGD modelling step in BRISQUE is not essential for final performance.
最近提出的盲/无参考图像空间质量评估器(BRISQUE)模型在盲图像质量评估(IQA)中实现了最先进的性能。该模型使用预定义的广义高斯分布(GGD)来描述自然场景统计的规律性,引入了由于图像内容变化而产生的拟合误差。本文提出了一个更广义的模型来更好地表征广泛的图像内容的规律性,该模型是通过相邻像素的平均减去对比度归一化(MSCN)系数和MSCN系数的成对乘积的连接直方图来学习的。基于MSCN的新特征显示了其保持图像统计量固有分布的能力。因此,支持向量机回归(SVR)可以将其映射到更准确的图像质量分数。实验结果表明,该方法从BRISQUE中获得了轻微的增益,这表明BRISQUE中精心制作的GGD建模步骤对最终性能不是必需的。
{"title":"Blind image quality assessment based on a new feature of nature scene statistics","authors":"Li Song, Chen Chen, Yi Xu, Genjian Xue, Yi Zhou","doi":"10.1109/VCIP.2014.7051498","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051498","url":null,"abstract":"A recently proposed model, known as blind/referenceless image spatial quality evaluator (BRISQUE), achieves the state-of-the-art performance in context of blind image quality assessment (IQA). This model used the predefined generalized Gaussian distribution (GGD) to describe the regularity of natural scene statistics, introducing fitting errors due to variations of image contents. In this paper, a more generalized model is proposed to better characterize the regularity of extensive image contents, which is learned from the concatenated histograms of mean subtracted contrast normalized (MSCN) coefficients and pairwise products of MSCN coefficients of neighbouring pixels. The new feature based on MSCN shows its capability of preserving intrinsic distribution of image statistics. Consequently support vector machine regression (SVR) can map it to more accurate image quality scores. Experimental results show that the proposed approach achieves a slight gain from BRISQUE, which indicates the crafted GGD modelling step in BRISQUE is not essential for final performance.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115904665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
DLP based anti-piracy display system 基于DLP的防盗版显示系统
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051525
Zhongpai Gao, Guangtao Zhai, Xiaolin Wu, Xiongkuo Min, Cheng Zhi
Camcorder piracy has great impact on the movie industry. Although there are many methods to prevent recording in theatre, no recognized technology satisfies the need of defeating camcorder piracy as well as having no effect on the audience. This paper presents a new projector display technique to defeat camcorder piracy in the theatre using a new paradigm of information display technology, called temporal psychovisual modulation (TPVM). TPVM exploits the difference in image formation mechanisms of human eyes and imaging sensors. The images formed in human vision is continuous integration of the light field while discrete sampling is used in digital video acquisition which has "blackout" period in each sampling cycle. Based on this difference, we can decompose a movie into a set of display frames and broadcast them out at high speed so that the audience can not notice any disturbance, while the video frames captured by camcorder will contain highly objectionable artifacts. The proposed prototype system built on the platform of DLP® LightCrafter 4500™ serves as a proof-of-concept of anti-piracy system.
摄像机盗版对电影行业的影响很大。虽然有许多方法可以防止在剧院录音,但没有一种公认的技术既能满足打击盗版摄像机的需要,又不会对观众产生影响。本文提出了一种新的投影仪显示技术,利用一种新的信息显示技术范式,称为时间心理视觉调制(TPVM),来打击剧院中的摄像机盗版。TPVM利用人眼和成像传感器在成像机制上的差异。人眼视觉中形成的图像是光场的连续积分,而数字视频采集采用离散采样,每个采样周期都有“停电”期。基于这种差异,我们可以将电影分解成一组显示帧,并以高速播放的方式将其播放出来,使观众察觉不到任何干扰,而摄像机捕捉到的视频帧将包含非常令人反感的伪影。拟议的原型系统建立在DLP®LightCrafter 4500™平台上,作为反盗版系统的概念验证。
{"title":"DLP based anti-piracy display system","authors":"Zhongpai Gao, Guangtao Zhai, Xiaolin Wu, Xiongkuo Min, Cheng Zhi","doi":"10.1109/VCIP.2014.7051525","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051525","url":null,"abstract":"Camcorder piracy has great impact on the movie industry. Although there are many methods to prevent recording in theatre, no recognized technology satisfies the need of defeating camcorder piracy as well as having no effect on the audience. This paper presents a new projector display technique to defeat camcorder piracy in the theatre using a new paradigm of information display technology, called temporal psychovisual modulation (TPVM). TPVM exploits the difference in image formation mechanisms of human eyes and imaging sensors. The images formed in human vision is continuous integration of the light field while discrete sampling is used in digital video acquisition which has \"blackout\" period in each sampling cycle. Based on this difference, we can decompose a movie into a set of display frames and broadcast them out at high speed so that the audience can not notice any disturbance, while the video frames captured by camcorder will contain highly objectionable artifacts. The proposed prototype system built on the platform of DLP® LightCrafter 4500™ serves as a proof-of-concept of anti-piracy system.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115699688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Fast multiple-view denoising based on image reconstruction by plane sweeping 基于平面扫描图像重构的快速多视点去噪
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051606
Mari Miyata, K. Kodama, T. Hamamoto
Denoising is important in image processing because degradation by noise affects not only the quality of captured images but also the performance of visual applications that use them. For example, under low light levels, it is difficult to accurately estimate scene depths using noisy stereo images. Conventional methods for denoising find similar regions on an image or among multiple images by block matching(BM) to integrate them for suppressing noise effectively. However, such exhaustive BM incurs considerable costs for real-time applications, in particular, when multi-view images(MVI) are involved. We use view-dependent plane sweeping(PS) for image reconstruction to achieve effective MVI denoising with low computational cost. We use PS for converting MVI to multi-focus images(MFI) to suppress their noise. Then, we find regions in focus on the MFI solely by comparing them with the target view image. Finally, we simply merge the regions to obtain reconstructed images in which their noise is effectively suppressed.
去噪在图像处理中很重要,因为噪声的退化不仅会影响捕获图像的质量,还会影响使用图像的视觉应用程序的性能。例如,在低光照水平下,使用有噪声的立体图像很难准确估计场景深度。传统的去噪方法是通过块匹配(BM)来寻找图像上或多幅图像之间的相似区域,并对其进行整合,从而有效地抑制噪声。然而,对于实时应用程序,特别是涉及多视图图像(MVI)时,这种详尽的BM会带来相当大的成本。我们使用视相关平面扫描(PS)进行图像重建,以较低的计算成本实现有效的MVI去噪。我们使用PS将MVI转换成多焦点图像(MFI),以抑制其噪声。然后,仅通过与目标视图图像的比较,我们就可以找到MFI上的焦点区域。最后,我们简单地合并区域以获得有效抑制噪声的重建图像。
{"title":"Fast multiple-view denoising based on image reconstruction by plane sweeping","authors":"Mari Miyata, K. Kodama, T. Hamamoto","doi":"10.1109/VCIP.2014.7051606","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051606","url":null,"abstract":"Denoising is important in image processing because degradation by noise affects not only the quality of captured images but also the performance of visual applications that use them. For example, under low light levels, it is difficult to accurately estimate scene depths using noisy stereo images. Conventional methods for denoising find similar regions on an image or among multiple images by block matching(BM) to integrate them for suppressing noise effectively. However, such exhaustive BM incurs considerable costs for real-time applications, in particular, when multi-view images(MVI) are involved. We use view-dependent plane sweeping(PS) for image reconstruction to achieve effective MVI denoising with low computational cost. We use PS for converting MVI to multi-focus images(MFI) to suppress their noise. Then, we find regions in focus on the MFI solely by comparing them with the target view image. Finally, we simply merge the regions to obtain reconstructed images in which their noise is effectively suppressed.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127401476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Simplified depth-based block partitioning and prediction merging in 3D video coding 简化3D视频编码中基于深度的块划分和预测合并
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051519
Fabian Jäger, M. Wien
3D video is an emerging technology that bundles depth information with texture videos to allow for view synthesis applications at the receiver. Depth discontinuities define object boundaries in both, depth maps and the collocated texture video. Therefore, depth segmentation can be utilized for a fine-grained motion field partitioning of the corresponding texture component. In this paper, depth information is used to increase coding efficiency for texture videos by deriving an arbitrarily shaped partitioning. By applying motion compensation to each partition independently and eventually merging the two prediction signals, highly accurate prediction signals can be produced that reduce the remaining texture residual signal significantly. Simulation results show bitrate savings of up to 2.8% for the dependent texture views and up to about 1.0% with respect to the total bitrate.
3D视频是一项新兴技术,它将深度信息与纹理视频捆绑在一起,允许接收器的视图合成应用。深度不连续在深度图和并置纹理视频中定义对象边界。因此,可以利用深度分割对相应纹理分量进行细粒度的运动场划分。本文利用深度信息对纹理视频进行任意形状的分割,提高编码效率。通过对每个分块分别进行运动补偿,最终将两个预测信号合并,可以产生高精度的预测信号,显著减少剩余纹理残留信号。模拟结果显示,对于依赖纹理视图,比特率节省高达2.8%,相对于总比特率节省高达约1.0%。
{"title":"Simplified depth-based block partitioning and prediction merging in 3D video coding","authors":"Fabian Jäger, M. Wien","doi":"10.1109/VCIP.2014.7051519","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051519","url":null,"abstract":"3D video is an emerging technology that bundles depth information with texture videos to allow for view synthesis applications at the receiver. Depth discontinuities define object boundaries in both, depth maps and the collocated texture video. Therefore, depth segmentation can be utilized for a fine-grained motion field partitioning of the corresponding texture component. In this paper, depth information is used to increase coding efficiency for texture videos by deriving an arbitrarily shaped partitioning. By applying motion compensation to each partition independently and eventually merging the two prediction signals, highly accurate prediction signals can be produced that reduce the remaining texture residual signal significantly. Simulation results show bitrate savings of up to 2.8% for the dependent texture views and up to about 1.0% with respect to the total bitrate.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116297947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2014 IEEE Visual Communications and Image Processing Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1