首页 > 最新文献

2014 IEEE Visual Communications and Image Processing Conference最新文献

英文 中文
Two-dimensional histogram expansion of wavelet coefficient for reversible data hiding 可逆数据隐藏的小波系数二维直方图展开
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051553
Kazuki Yamato, Kazuma Shinoda, Madoka Hasegawa, Shigeo Kato
In this paper, we propose a reversible data hiding (RDH) method based on a two-dimensional wavelet coefficient histogram (2D WCH) in the wavelet domain. First, a cover image is decomposed into wavelet subbands using the invertible integer-to-integer wavelet transform (121-WT). Then, the 2D WCH is generated by counting the occurrence frequency of the wavelet coefficient pairs which denote two wavelet coefficients located in the same position in the selected two subbands where the secret message is embedded. By using the 2D WCH, the correlation between the selected subbands is more effectively utilized than the traditional ID histogram. The proposed method embed the secret message reversibly in the cover image by expanding the 2D WCH. In order to embed the secret message as efficient as possible, the expansion rule for 2D WCH is proposed. Moreover, the coefficient pair selection (CPS), which the coefficients embedding the data are selected in order to modify only the selected coefficients, is implemented before generating the 2D WCH. In the experiment, the proposed method is compared with the conventional RDH methods in terms of the capacity-distortion curve.
本文提出了一种基于小波域二维小波系数直方图(2D WCH)的可逆数据隐藏方法。首先,利用可逆整数到整数小波变换(121-WT)将封面图像分解成小波子带;然后,通过计算小波系数对的出现频率来生成二维WCH,这些小波系数对表示在选定的嵌入秘密信息的两个子带中位于相同位置的两个小波系数。与传统的ID直方图相比,使用二维WCH可以更有效地利用所选子带之间的相关性。该方法通过扩展二维WCH,将秘密信息可逆嵌入到封面图像中。为了尽可能高效地嵌入秘密信息,提出了二维WCH的展开规则。在生成二维WCH之前,实现了系数对选择(CPS),即选择嵌入数据的系数,只修改被选中的系数。在实验中,将该方法与传统的RDH方法在容量畸变曲线上进行了比较。
{"title":"Two-dimensional histogram expansion of wavelet coefficient for reversible data hiding","authors":"Kazuki Yamato, Kazuma Shinoda, Madoka Hasegawa, Shigeo Kato","doi":"10.1109/VCIP.2014.7051553","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051553","url":null,"abstract":"In this paper, we propose a reversible data hiding (RDH) method based on a two-dimensional wavelet coefficient histogram (2D WCH) in the wavelet domain. First, a cover image is decomposed into wavelet subbands using the invertible integer-to-integer wavelet transform (121-WT). Then, the 2D WCH is generated by counting the occurrence frequency of the wavelet coefficient pairs which denote two wavelet coefficients located in the same position in the selected two subbands where the secret message is embedded. By using the 2D WCH, the correlation between the selected subbands is more effectively utilized than the traditional ID histogram. The proposed method embed the secret message reversibly in the cover image by expanding the 2D WCH. In order to embed the secret message as efficient as possible, the expansion rule for 2D WCH is proposed. Moreover, the coefficient pair selection (CPS), which the coefficients embedding the data are selected in order to modify only the selected coefficients, is implemented before generating the 2D WCH. In the experiment, the proposed method is compared with the conventional RDH methods in terms of the capacity-distortion curve.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121153626","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Robust 3D LUT estimation method for SHVC color gamut scalability SHVC色域可扩展性的鲁棒3D LUT估计方法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051510
Yuwen He, Yan Ye, Jie Dong
Color gamut scalability (CGS) in scalable extensions of High Efficiency Video Coding (SHVC) supports scalable coding with multiple layers in different color spaces. Base layer conveying HDTV video in BT.709 color space and enhancement layer conveying UHDTV video in BT.2020 color space is identified as a practical use case for CGS. Efficient CGS coding can be achieved using a 3D Look-up Table (LUT) based color conversion process. This paper proposes a robust 3D LUT parameter estimation method that estimates the 3D LUT parameters globally using the Least Square method. Problems of matrix sparsity and uneven sample distribution are carefully handled to improve the stability and accuracy of the estimation process. Simulation results confirm that the proposed 3D LUT estimation method can significantly improve coding performance compared with other gamut conversion methods.
高效视频编码(High Efficiency Video Coding, SHVC)的可扩展扩展中的色域可扩展性(CGS)支持在不同颜色空间中进行多层可扩展编码。确定了在BT.709色彩空间中传输HDTV视频的基层和在BT.2020色彩空间中传输UHDTV视频的增强层作为CGS的实际用例。使用基于3D查找表(LUT)的颜色转换过程可以实现高效的CGS编码。提出了一种鲁棒的三维LUT参数估计方法,利用最小二乘法对三维LUT参数进行全局估计。为了提高估计过程的稳定性和准确性,对矩阵稀疏性和样本分布不均匀等问题进行了细致的处理。仿真结果表明,与其他色域转换方法相比,所提出的三维LUT估计方法可以显著提高编码性能。
{"title":"Robust 3D LUT estimation method for SHVC color gamut scalability","authors":"Yuwen He, Yan Ye, Jie Dong","doi":"10.1109/VCIP.2014.7051510","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051510","url":null,"abstract":"Color gamut scalability (CGS) in scalable extensions of High Efficiency Video Coding (SHVC) supports scalable coding with multiple layers in different color spaces. Base layer conveying HDTV video in BT.709 color space and enhancement layer conveying UHDTV video in BT.2020 color space is identified as a practical use case for CGS. Efficient CGS coding can be achieved using a 3D Look-up Table (LUT) based color conversion process. This paper proposes a robust 3D LUT parameter estimation method that estimates the 3D LUT parameters globally using the Least Square method. Problems of matrix sparsity and uneven sample distribution are carefully handled to improve the stability and accuracy of the estimation process. Simulation results confirm that the proposed 3D LUT estimation method can significantly improve coding performance compared with other gamut conversion methods.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115380796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Blind image quality assessment based on a new feature of nature scene statistics 基于自然场景统计新特征的盲图像质量评价
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051498
Li Song, Chen Chen, Yi Xu, Genjian Xue, Yi Zhou
A recently proposed model, known as blind/referenceless image spatial quality evaluator (BRISQUE), achieves the state-of-the-art performance in context of blind image quality assessment (IQA). This model used the predefined generalized Gaussian distribution (GGD) to describe the regularity of natural scene statistics, introducing fitting errors due to variations of image contents. In this paper, a more generalized model is proposed to better characterize the regularity of extensive image contents, which is learned from the concatenated histograms of mean subtracted contrast normalized (MSCN) coefficients and pairwise products of MSCN coefficients of neighbouring pixels. The new feature based on MSCN shows its capability of preserving intrinsic distribution of image statistics. Consequently support vector machine regression (SVR) can map it to more accurate image quality scores. Experimental results show that the proposed approach achieves a slight gain from BRISQUE, which indicates the crafted GGD modelling step in BRISQUE is not essential for final performance.
最近提出的盲/无参考图像空间质量评估器(BRISQUE)模型在盲图像质量评估(IQA)中实现了最先进的性能。该模型使用预定义的广义高斯分布(GGD)来描述自然场景统计的规律性,引入了由于图像内容变化而产生的拟合误差。本文提出了一个更广义的模型来更好地表征广泛的图像内容的规律性,该模型是通过相邻像素的平均减去对比度归一化(MSCN)系数和MSCN系数的成对乘积的连接直方图来学习的。基于MSCN的新特征显示了其保持图像统计量固有分布的能力。因此,支持向量机回归(SVR)可以将其映射到更准确的图像质量分数。实验结果表明,该方法从BRISQUE中获得了轻微的增益,这表明BRISQUE中精心制作的GGD建模步骤对最终性能不是必需的。
{"title":"Blind image quality assessment based on a new feature of nature scene statistics","authors":"Li Song, Chen Chen, Yi Xu, Genjian Xue, Yi Zhou","doi":"10.1109/VCIP.2014.7051498","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051498","url":null,"abstract":"A recently proposed model, known as blind/referenceless image spatial quality evaluator (BRISQUE), achieves the state-of-the-art performance in context of blind image quality assessment (IQA). This model used the predefined generalized Gaussian distribution (GGD) to describe the regularity of natural scene statistics, introducing fitting errors due to variations of image contents. In this paper, a more generalized model is proposed to better characterize the regularity of extensive image contents, which is learned from the concatenated histograms of mean subtracted contrast normalized (MSCN) coefficients and pairwise products of MSCN coefficients of neighbouring pixels. The new feature based on MSCN shows its capability of preserving intrinsic distribution of image statistics. Consequently support vector machine regression (SVR) can map it to more accurate image quality scores. Experimental results show that the proposed approach achieves a slight gain from BRISQUE, which indicates the crafted GGD modelling step in BRISQUE is not essential for final performance.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115904665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Simplified depth-based block partitioning and prediction merging in 3D video coding 简化3D视频编码中基于深度的块划分和预测合并
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051519
Fabian Jäger, M. Wien
3D video is an emerging technology that bundles depth information with texture videos to allow for view synthesis applications at the receiver. Depth discontinuities define object boundaries in both, depth maps and the collocated texture video. Therefore, depth segmentation can be utilized for a fine-grained motion field partitioning of the corresponding texture component. In this paper, depth information is used to increase coding efficiency for texture videos by deriving an arbitrarily shaped partitioning. By applying motion compensation to each partition independently and eventually merging the two prediction signals, highly accurate prediction signals can be produced that reduce the remaining texture residual signal significantly. Simulation results show bitrate savings of up to 2.8% for the dependent texture views and up to about 1.0% with respect to the total bitrate.
3D视频是一项新兴技术,它将深度信息与纹理视频捆绑在一起,允许接收器的视图合成应用。深度不连续在深度图和并置纹理视频中定义对象边界。因此,可以利用深度分割对相应纹理分量进行细粒度的运动场划分。本文利用深度信息对纹理视频进行任意形状的分割,提高编码效率。通过对每个分块分别进行运动补偿,最终将两个预测信号合并,可以产生高精度的预测信号,显著减少剩余纹理残留信号。模拟结果显示,对于依赖纹理视图,比特率节省高达2.8%,相对于总比特率节省高达约1.0%。
{"title":"Simplified depth-based block partitioning and prediction merging in 3D video coding","authors":"Fabian Jäger, M. Wien","doi":"10.1109/VCIP.2014.7051519","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051519","url":null,"abstract":"3D video is an emerging technology that bundles depth information with texture videos to allow for view synthesis applications at the receiver. Depth discontinuities define object boundaries in both, depth maps and the collocated texture video. Therefore, depth segmentation can be utilized for a fine-grained motion field partitioning of the corresponding texture component. In this paper, depth information is used to increase coding efficiency for texture videos by deriving an arbitrarily shaped partitioning. By applying motion compensation to each partition independently and eventually merging the two prediction signals, highly accurate prediction signals can be produced that reduce the remaining texture residual signal significantly. Simulation results show bitrate savings of up to 2.8% for the dependent texture views and up to about 1.0% with respect to the total bitrate.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116297947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DLP based anti-piracy display system 基于DLP的防盗版显示系统
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051525
Zhongpai Gao, Guangtao Zhai, Xiaolin Wu, Xiongkuo Min, Cheng Zhi
Camcorder piracy has great impact on the movie industry. Although there are many methods to prevent recording in theatre, no recognized technology satisfies the need of defeating camcorder piracy as well as having no effect on the audience. This paper presents a new projector display technique to defeat camcorder piracy in the theatre using a new paradigm of information display technology, called temporal psychovisual modulation (TPVM). TPVM exploits the difference in image formation mechanisms of human eyes and imaging sensors. The images formed in human vision is continuous integration of the light field while discrete sampling is used in digital video acquisition which has "blackout" period in each sampling cycle. Based on this difference, we can decompose a movie into a set of display frames and broadcast them out at high speed so that the audience can not notice any disturbance, while the video frames captured by camcorder will contain highly objectionable artifacts. The proposed prototype system built on the platform of DLP® LightCrafter 4500™ serves as a proof-of-concept of anti-piracy system.
摄像机盗版对电影行业的影响很大。虽然有许多方法可以防止在剧院录音,但没有一种公认的技术既能满足打击盗版摄像机的需要,又不会对观众产生影响。本文提出了一种新的投影仪显示技术,利用一种新的信息显示技术范式,称为时间心理视觉调制(TPVM),来打击剧院中的摄像机盗版。TPVM利用人眼和成像传感器在成像机制上的差异。人眼视觉中形成的图像是光场的连续积分,而数字视频采集采用离散采样,每个采样周期都有“停电”期。基于这种差异,我们可以将电影分解成一组显示帧,并以高速播放的方式将其播放出来,使观众察觉不到任何干扰,而摄像机捕捉到的视频帧将包含非常令人反感的伪影。拟议的原型系统建立在DLP®LightCrafter 4500™平台上,作为反盗版系统的概念验证。
{"title":"DLP based anti-piracy display system","authors":"Zhongpai Gao, Guangtao Zhai, Xiaolin Wu, Xiongkuo Min, Cheng Zhi","doi":"10.1109/VCIP.2014.7051525","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051525","url":null,"abstract":"Camcorder piracy has great impact on the movie industry. Although there are many methods to prevent recording in theatre, no recognized technology satisfies the need of defeating camcorder piracy as well as having no effect on the audience. This paper presents a new projector display technique to defeat camcorder piracy in the theatre using a new paradigm of information display technology, called temporal psychovisual modulation (TPVM). TPVM exploits the difference in image formation mechanisms of human eyes and imaging sensors. The images formed in human vision is continuous integration of the light field while discrete sampling is used in digital video acquisition which has \"blackout\" period in each sampling cycle. Based on this difference, we can decompose a movie into a set of display frames and broadcast them out at high speed so that the audience can not notice any disturbance, while the video frames captured by camcorder will contain highly objectionable artifacts. The proposed prototype system built on the platform of DLP® LightCrafter 4500™ serves as a proof-of-concept of anti-piracy system.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115699688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Effective H.264/AVC to HEVC transcoder based on prediction homogeneity 基于预测同质性的高效H.264/AVC转HEVC转码器
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051547
Feiyang Zheng, Zhiru Shi, Xiaoyun Zhang, Zhiyong Gao
The new video coding standard, High Efficiency Video Coding (HEVC), has been established to succeed the widely used H.264/AVC standard. However, an enormous amount of legacy content is encoded with H.264/AVC. This makes high performance AVC to HEVC transcoding in great need. This paper presents a fast transcoding algorithm based on residual and motion information extracted from H.264 decoder. By exploiting these side information, regions' homogeneity characteristic are analysed. An efficient coding unit (CU) and prediction unit (PU) mode decision strategy is proposed combing regions' prediction homogeneity and current encoding information. The experimental results show that the proposed transcoding scheme can save up to 55% of encoding time with negligible loss of coding efficiency, when compared to that of the full decoding and full encoding transcoder.
高效视频编码(High Efficiency video coding, HEVC)是为了取代目前广泛使用的H.264/AVC标准而建立的一种新的视频编码标准。然而,大量的遗留内容是用H.264/AVC编码的。这使得高性能的AVC到HEVC转码的需求非常大。提出了一种基于H.264解码器残差和运动信息提取的快速转码算法。利用这些侧信息,分析了区域的均匀性特征。结合区域预测同质性和当前编码信息,提出了一种高效的编码单元和预测单元模式决策策略。实验结果表明,与全解码和全编码转码器相比,所提出的转码方案可节省高达55%的编码时间,而编码效率的损失可以忽略不计。
{"title":"Effective H.264/AVC to HEVC transcoder based on prediction homogeneity","authors":"Feiyang Zheng, Zhiru Shi, Xiaoyun Zhang, Zhiyong Gao","doi":"10.1109/VCIP.2014.7051547","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051547","url":null,"abstract":"The new video coding standard, High Efficiency Video Coding (HEVC), has been established to succeed the widely used H.264/AVC standard. However, an enormous amount of legacy content is encoded with H.264/AVC. This makes high performance AVC to HEVC transcoding in great need. This paper presents a fast transcoding algorithm based on residual and motion information extracted from H.264 decoder. By exploiting these side information, regions' homogeneity characteristic are analysed. An efficient coding unit (CU) and prediction unit (PU) mode decision strategy is proposed combing regions' prediction homogeneity and current encoding information. The experimental results show that the proposed transcoding scheme can save up to 55% of encoding time with negligible loss of coding efficiency, when compared to that of the full decoding and full encoding transcoder.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"80 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115012299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Probabilistic chunk scheduling approach in parallel multiple-server DASH 并行多服务器DASH中的概率块调度方法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051490
Li Liu, Chao Zhou, Xinggong Zhang, Zongming Guo, Cheng Li
Recently parallel Dynamic Adaptive Streaming over HTTP (DASH) has emerged as a promising way to supply higher bandwidth, connection diversity and reliability. However, it is still a big challenge to download chunks sequentially in parallel DASH due to heterogeneous and time-varying bandwidth of multiple servers. In this paper, we propose a novel probabilistic chunk scheduling approach considering time-varying bandwidth. Video chunks are scheduled to the servers which consume the least time while with the highest probability to complete downloading before the deadline. The proposed approach is formulated as a constrained optimization problem with the objective to minimize the total downloading time. Using the probabilistic model of time-varying bandwidth, we first estimate the probability of successful downloading chunks before the playback deadline. Then we estimate the download time of chunks. A near-optimal solution algorithm is designed which schedules chunks to the servers with minimal downloading time while the completion probability is under the constraint. Compared with the existing schemes, the experimental results demonstrate that our proposed scheme greatly increases the number of chunks that are received orderly.
最近,并行动态自适应HTTP流(DASH)作为一种提供更高带宽、连接多样性和可靠性的有前途的方式出现了。然而,由于多台服务器的异构和时变带宽,在并行DASH中按顺序下载数据块仍然是一个很大的挑战。本文提出了一种考虑时变带宽的概率块调度方法。视频块被安排到消耗时间最少的服务器上,而在截止日期之前完成下载的可能性最大。该方法是一个以最小化总下载时间为目标的约束优化问题。利用时变带宽的概率模型,我们首先估计了在播放截止日期之前成功下载数据块的概率。然后估计块的下载时间。设计了一种近乎最优的算法,在完成概率受限的情况下,以最小的下载时间将数据块调度到服务器上。实验结果表明,与现有方案相比,我们提出的方案大大增加了有序接收块的数量。
{"title":"Probabilistic chunk scheduling approach in parallel multiple-server DASH","authors":"Li Liu, Chao Zhou, Xinggong Zhang, Zongming Guo, Cheng Li","doi":"10.1109/VCIP.2014.7051490","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051490","url":null,"abstract":"Recently parallel Dynamic Adaptive Streaming over HTTP (DASH) has emerged as a promising way to supply higher bandwidth, connection diversity and reliability. However, it is still a big challenge to download chunks sequentially in parallel DASH due to heterogeneous and time-varying bandwidth of multiple servers. In this paper, we propose a novel probabilistic chunk scheduling approach considering time-varying bandwidth. Video chunks are scheduled to the servers which consume the least time while with the highest probability to complete downloading before the deadline. The proposed approach is formulated as a constrained optimization problem with the objective to minimize the total downloading time. Using the probabilistic model of time-varying bandwidth, we first estimate the probability of successful downloading chunks before the playback deadline. Then we estimate the download time of chunks. A near-optimal solution algorithm is designed which schedules chunks to the servers with minimal downloading time while the completion probability is under the constraint. Compared with the existing schemes, the experimental results demonstrate that our proposed scheme greatly increases the number of chunks that are received orderly.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129422509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Fast CU splitting and pruning method based on online learning for intra coding in HEVC 基于在线学习的HEVC内编码快速CU分割与剪枝方法
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051603
Yuchen Li, Yitong Liu, Hongwen Yang, Dacheng Yang
The latest standard of High Efficiency Video Coding (HEVC) has a better coding efficiency compared to H.264/AVC. It is reported that the bitrate of the video applied HEVC is the half of video applied H.264/AVC at the same encoding quality. However, the cost of improvement is the increasing computational complexity which is mainly brought by the quadtree based coding tree unit (CTU). In this paper, a fast coding units (CU) splitting and pruning method is proposed to speed up the process of searching the best partition for CTU. Experiment has shown that our method can save 46% computational complexity on average at the cost of increasing Bjontegaard delta rate (BD-rate) by 0.82% when the method is applied to sequences in Class A.
与H.264/AVC相比,最新的高效视频编码(HEVC)标准具有更高的编码效率。据报道,在相同的编码质量下,采用HEVC的视频的比特率是采用H.264/AVC的视频的一半。然而,改进的代价是计算复杂度的增加,这主要是由基于四叉树的编码树单元(CTU)带来的。本文提出了一种快速编码单元分割和剪枝方法,以加快搜索CTU最佳分区的速度。实验表明,当该方法应用于A类序列时,以提高Bjontegaard δ率(BD-rate) 0.82%为代价,平均节省了46%的计算复杂度。
{"title":"Fast CU splitting and pruning method based on online learning for intra coding in HEVC","authors":"Yuchen Li, Yitong Liu, Hongwen Yang, Dacheng Yang","doi":"10.1109/VCIP.2014.7051603","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051603","url":null,"abstract":"The latest standard of High Efficiency Video Coding (HEVC) has a better coding efficiency compared to H.264/AVC. It is reported that the bitrate of the video applied HEVC is the half of video applied H.264/AVC at the same encoding quality. However, the cost of improvement is the increasing computational complexity which is mainly brought by the quadtree based coding tree unit (CTU). In this paper, a fast coding units (CU) splitting and pruning method is proposed to speed up the process of searching the best partition for CTU. Experiment has shown that our method can save 46% computational complexity on average at the cost of increasing Bjontegaard delta rate (BD-rate) by 0.82% when the method is applied to sequences in Class A.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129422541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Video compressive sensing via structured Laplacian modelling 基于结构化拉普拉斯建模的视频压缩感知
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051591
Chen Zhao, Siwei Ma, Wen Gao
Seeking a fair domain in which the signal can exhibit high sparsity is of essential significance in compressive sensing (CS). Most methods in the literature, however, use a fixed transform domain or prior information, which cannot adapt to various video contents. In this paper, we propose a video CS recovery algorithm based on the structured Laplacian model, which can effectually deal with the non-stationarity of natural videos. To build the model, structured patch groups are constructed according to the nonlocal similarity in a temporal scope. By incorporating the model into the CS paradigm, we can formulate an ℓ1-norm optimization problem, for which a solution based on the iterative shrinkage/thresholding algorithms (ISTA) is designed. Experimental results demonstrate that the proposed algorithm outperforms the state-of-the-art methods in both objective and subjective recovery quality.
在压缩感知(CS)中,寻找一个信号能表现出高稀疏性的公平域是至关重要的。然而,文献中的大多数方法使用固定的变换域或先验信息,无法适应各种视频内容。本文提出了一种基于结构化拉普拉斯模型的视频CS恢复算法,该算法可以有效地处理自然视频的非平稳性。为了构建模型,根据时间范围内的非局部相似度构造结构化补丁群。通过将该模型纳入CS范式,我们可以提出一个1-范数优化问题,并设计了基于迭代收缩/阈值算法(ISTA)的解决方案。实验结果表明,该算法在客观和主观恢复质量上都优于现有方法。
{"title":"Video compressive sensing via structured Laplacian modelling","authors":"Chen Zhao, Siwei Ma, Wen Gao","doi":"10.1109/VCIP.2014.7051591","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051591","url":null,"abstract":"Seeking a fair domain in which the signal can exhibit high sparsity is of essential significance in compressive sensing (CS). Most methods in the literature, however, use a fixed transform domain or prior information, which cannot adapt to various video contents. In this paper, we propose a video CS recovery algorithm based on the structured Laplacian model, which can effectually deal with the non-stationarity of natural videos. To build the model, structured patch groups are constructed according to the nonlocal similarity in a temporal scope. By incorporating the model into the CS paradigm, we can formulate an ℓ1-norm optimization problem, for which a solution based on the iterative shrinkage/thresholding algorithms (ISTA) is designed. Experimental results demonstrate that the proposed algorithm outperforms the state-of-the-art methods in both objective and subjective recovery quality.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"117 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130217876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Depth coding using depth discontinuity prediction and in-loop boundary reconstruction filtering 使用深度不连续预测和环内边界重建滤波的深度编码
Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051597
R. Farrugia, Maverick Hili
This paper presents a depth coding strategy that employs K-means clustering to segment the sequence of depth images into K clusters. The resulting clusters are losslessly compressed and transmitted as supplemental enhancement information to aid the decoder in predicting macroblocks containing depth discontinuities. This method further employs an in-loop boundary reconstruction filter to reduce distortions at the edges. The proposed algorithm was integrated within both H.264/AVC and H.264/MVC video coding standards. Simulation results demonstrate that the proposed scheme outperforms the state of the art depth coding schemes, where rendered Peak Signal to Noise Ratio (PSNR) gains between 0.1 dB and 0.5 dB were observed.
本文提出了一种深度编码策略,利用K均值聚类将深度图像序列分割成K个聚类。所得到的聚类被无损压缩并作为补充增强信息传输,以帮助解码器预测包含深度不连续的宏块。该方法进一步采用环内边界重建滤波器来减少边缘处的畸变。该算法集成在H.264/AVC和H.264/MVC视频编码标准中。仿真结果表明,所提出的方案优于当前深度编码方案,其中呈现的峰值信噪比(PSNR)增益在0.1 dB和0.5 dB之间。
{"title":"Depth coding using depth discontinuity prediction and in-loop boundary reconstruction filtering","authors":"R. Farrugia, Maverick Hili","doi":"10.1109/VCIP.2014.7051597","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051597","url":null,"abstract":"This paper presents a depth coding strategy that employs K-means clustering to segment the sequence of depth images into K clusters. The resulting clusters are losslessly compressed and transmitted as supplemental enhancement information to aid the decoder in predicting macroblocks containing depth discontinuities. This method further employs an in-loop boundary reconstruction filter to reduce distortions at the edges. The proposed algorithm was integrated within both H.264/AVC and H.264/MVC video coding standards. Simulation results demonstrate that the proposed scheme outperforms the state of the art depth coding schemes, where rendered Peak Signal to Noise Ratio (PSNR) gains between 0.1 dB and 0.5 dB were observed.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131030883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2014 IEEE Visual Communications and Image Processing Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1