首页 > 最新文献

28th Picture Coding Symposium最新文献

英文 中文
Influences of frame delay and packet loss between left and right frames in stereoscopic video communications 立体视频通信中左右帧间帧延迟和丢包的影响
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702550
Shuliang Lin, Yuichiro Sawa, Norishige Fukushima, Y. Ishibashi
This paper analyzes the influences of frame delay and packet loss on stereoscopic vision when stereoscopic video is transferred over a IP network. We employ live action videos which are transferred to a head-mount-display (HMD) and do the assessment on stereoscopic perception. As a result, we found that speed and movement direction of the attention object play a great role on the deterioration when frame delay and packet loss exist.
本文分析了在IP网络上传输立体视频时,帧延迟和丢包对立体视觉的影响。我们采用实景视频传输到头戴式显示器(HMD),并对立体感知进行评估。结果发现,当存在帧延迟和丢包时,注意对象的运动速度和运动方向对图像的劣化有很大影响。
{"title":"Influences of frame delay and packet loss between left and right frames in stereoscopic video communications","authors":"Shuliang Lin, Yuichiro Sawa, Norishige Fukushima, Y. Ishibashi","doi":"10.1109/PCS.2010.5702550","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702550","url":null,"abstract":"This paper analyzes the influences of frame delay and packet loss on stereoscopic vision when stereoscopic video is transferred over a IP network. We employ live action videos which are transferred to a head-mount-display (HMD) and do the assessment on stereoscopic perception. As a result, we found that speed and movement direction of the attention object play a great role on the deterioration when frame delay and packet loss exist.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115777503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On-line statistical analysis based fast mode decision for multi-view video coding 基于在线统计分析的多视点视频编码快速模式决策
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702541
G. Chan, Jheng-Ping Lin, A. Tang
The high computational complexity of multi-view video codecs makes it necessary to speed up for their realization in consumer electronics. Since fast encoding algorithms are expected to adapt to different video sequences, this paper proposes a fast algorithm that consists of fast mode decision and fast disparity estimation for multi-view video coding. The fast mode decision algorithm applies to both temporal and inter-view predictions. The candidates for mode decision are reduced based on a set of thresholds. Differ from the previous fast mode decision algorithms for MVC, this scheme determines the thresholds according to the online statistical analysis of motion and disparity costs of the first GOP in each view. Since the inter-view prediction is time consuming, we propose a fast disparity estimation algorithm to save encoding time. Experimental results show that our proposed scheme reduces the computational complexity significantly with negligible degradation of coding efficiency.
多视点视频编解码器的高计算复杂度使得其在消费类电子产品中的实现速度有必要加快。针对快速编码算法需要适应不同视频序列的特点,本文提出了一种由快速模式判定和快速视差估计组成的多视点视频编码快速算法。快速模式决策算法既适用于时间预测,也适用于跨视图预测。根据一组阈值减少模式决策的候选项。与以往的MVC快速模式决策算法不同,该方案通过对每个视图中第一个GOP的运动和视差代价的在线统计分析来确定阈值。针对视差预测耗时的问题,提出了一种快速视差估计算法,以节省编码时间。实验结果表明,该方案在降低编码效率的基础上显著降低了计算复杂度。
{"title":"On-line statistical analysis based fast mode decision for multi-view video coding","authors":"G. Chan, Jheng-Ping Lin, A. Tang","doi":"10.1109/PCS.2010.5702541","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702541","url":null,"abstract":"The high computational complexity of multi-view video codecs makes it necessary to speed up for their realization in consumer electronics. Since fast encoding algorithms are expected to adapt to different video sequences, this paper proposes a fast algorithm that consists of fast mode decision and fast disparity estimation for multi-view video coding. The fast mode decision algorithm applies to both temporal and inter-view predictions. The candidates for mode decision are reduced based on a set of thresholds. Differ from the previous fast mode decision algorithms for MVC, this scheme determines the thresholds according to the online statistical analysis of motion and disparity costs of the first GOP in each view. Since the inter-view prediction is time consuming, we propose a fast disparity estimation algorithm to save encoding time. Experimental results show that our proposed scheme reduces the computational complexity significantly with negligible degradation of coding efficiency.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128866064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Super-resolution decoding of JPEG-compressed image data with the shrinkage in the redundant DCT domain 利用冗余DCT域的收缩对jpeg压缩图像数据进行超分辨率解码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702436
T. Komatsu, Yasutaka Ueda, T. Saito
Alter, Durand and Froment introduced the total-variation (TV) minimization approach to the artifact-free JPEG decoding, which is referred to as the ADF decoding method [1]. They formulated the decoding problem as the constrained TV restoration problem, in which the TV seminorm of its restored color image is minimized under the constraint that each DCT coefficient of the restored color image should be in the quantization interval of its corresponding DCT coefficient of the JPEG-compressed data. This paper proposes a new restoration approach to the JPEG decoding. Instead of the TV regularization, our new JPEG-decoding method employs a shrinkage operation in the redundant DCT domain, to mitigate degradations caused by the JPEG coding. Our new method not only can selectively suppress ringing artifacts near color edges, but also can efficiently eliminate blocking artifacts in originally smoothly-varying image regions, where the blocking artifacts are very noticeable. Through decoding simulations, we experimentally show that our new decoding method can reduce JPEG-coding artifacts more effectively than the ADF decoding method.
Alter, Durand and Froment引入了全变差(total-variation, TV)最小化方法来实现无伪影的JPEG解码,称为ADF解码方法[1]。他们将解码问题表述为约束电视还原问题,即在还原的彩色图像的每个DCT系数应在jpeg压缩数据的相应DCT系数的量化区间内的约束下,将其还原的彩色图像的电视半形态最小化。本文提出了一种新的JPEG解码复原方法。我们的新JPEG解码方法采用冗余DCT域的收缩操作来代替TV正则化,以减轻JPEG编码引起的降级。该方法不仅可以选择性地抑制颜色边缘附近的环状伪影,而且可以有效地消除原本平滑变化的图像区域的阻塞伪影,在这些区域阻塞伪影非常明显。通过解码仿真,我们实验证明了我们的解码方法能比ADF解码方法更有效地减少jpeg编码伪影。
{"title":"Super-resolution decoding of JPEG-compressed image data with the shrinkage in the redundant DCT domain","authors":"T. Komatsu, Yasutaka Ueda, T. Saito","doi":"10.1109/PCS.2010.5702436","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702436","url":null,"abstract":"Alter, Durand and Froment introduced the total-variation (TV) minimization approach to the artifact-free JPEG decoding, which is referred to as the ADF decoding method [1]. They formulated the decoding problem as the constrained TV restoration problem, in which the TV seminorm of its restored color image is minimized under the constraint that each DCT coefficient of the restored color image should be in the quantization interval of its corresponding DCT coefficient of the JPEG-compressed data. This paper proposes a new restoration approach to the JPEG decoding. Instead of the TV regularization, our new JPEG-decoding method employs a shrinkage operation in the redundant DCT domain, to mitigate degradations caused by the JPEG coding. Our new method not only can selectively suppress ringing artifacts near color edges, but also can efficiently eliminate blocking artifacts in originally smoothly-varying image regions, where the blocking artifacts are very noticeable. Through decoding simulations, we experimentally show that our new decoding method can reduce JPEG-coding artifacts more effectively than the ADF decoding method.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"224 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131262123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Compressive video sensing based on user attention model 基于用户注意力模型的压缩视频感知
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702586
Jie Xu, Jianwei Ma, Dongming Zhang, Yongdong Zhang, Shouxun Lin
We propose a compressive video sensing scheme based on user attention model (UAM) for real video sequences acquisition. In this work, for every group of consecutive video frames, we set the first frame as reference frame and build a UAM with visual rhythm analysis (VRA) to automatically determine region-of-interest (ROI) for non-reference frames. The determined ROI usually has significant movement and attracts more attention. Each frame of the video sequence is divided into non-overlapping blocks of 16×16 pixel size. Compressive video sampling is conducted in a block-by-block manner on each frame through a single operator and in a whole region manner on the ROIs through a different operator. Our video reconstruction algorithm involves alternating direction l1 — norm minimization algorithm (ADM) for the frame difference of non-ROI blocks and minimum total-variance (TV) method for the ROIs. Experimental results showed that our method could significantly enhance the quality of reconstructed video and reduce the errors accumulated during the reconstruction.
提出了一种基于用户注意力模型(UAM)的压缩视频感知方案,用于真实视频序列的获取。在这项工作中,我们将每组连续视频帧的第一帧设置为参考帧,并构建带有视觉节奏分析(VRA)的UAM,以自动确定非参考帧的感兴趣区域(ROI)。确定的投资回报率通常具有较大的变动,引起较多的关注。视频序列的每一帧被分成不重叠的块,像素大小为16×16。压缩视频采样通过单个算子对每帧进行逐块采样,通过不同算子对roi进行全区域采样。我们的视频重建算法采用交替方向l1范数最小化算法(ADM)处理非roi块的帧差,最小总方差(TV)方法处理roi块的帧差。实验结果表明,该方法可以显著提高重构视频的质量,减少重构过程中积累的误差。
{"title":"Compressive video sensing based on user attention model","authors":"Jie Xu, Jianwei Ma, Dongming Zhang, Yongdong Zhang, Shouxun Lin","doi":"10.1109/PCS.2010.5702586","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702586","url":null,"abstract":"We propose a compressive video sensing scheme based on user attention model (UAM) for real video sequences acquisition. In this work, for every group of consecutive video frames, we set the first frame as reference frame and build a UAM with visual rhythm analysis (VRA) to automatically determine region-of-interest (ROI) for non-reference frames. The determined ROI usually has significant movement and attracts more attention. Each frame of the video sequence is divided into non-overlapping blocks of 16×16 pixel size. Compressive video sampling is conducted in a block-by-block manner on each frame through a single operator and in a whole region manner on the ROIs through a different operator. Our video reconstruction algorithm involves alternating direction l1 — norm minimization algorithm (ADM) for the frame difference of non-ROI blocks and minimum total-variance (TV) method for the ROIs. Experimental results showed that our method could significantly enhance the quality of reconstructed video and reduce the errors accumulated during the reconstruction.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115708671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Content-based retrieval by multiple image examples for sign board retrieval 基于内容的多图像实例检索标志板
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702447
A. Yoshitaka, Terumasa Hyoudou
In the area of retrieving image databases, one of the promising approaches is to retrieve it by specifying image example. However, specifying a single image example is not always sufficient to get satisfactory result, since one image example does not give comprehensive ranges of values that reflect the various aspects of the object to be retrieved. In this paper, we propose a method of retrieving images by specifying multiple image examples that is designed for retrieving sign boards. Features of color, shape, and spatial relation of color regions are extracted from example images, and they are clustered so as to obtain proper range of values. Compared with QBE systems that accept only a single image as the query condition, MIERS (Multi-Image Example-based Retrieval System) returns better retrieval result, where the experimental result showed that specifying more examples helps to improve recall with little deterioration of precision.
在图像数据库检索领域,通过指定图像样本进行检索是一种很有前途的方法。然而,指定单个图像示例并不总是足以获得令人满意的结果,因为一个图像示例不能给出反映要检索对象的各个方面的全面值范围。在本文中,我们提出了一种通过指定多个图像示例来检索图像的方法,该方法专为检索标牌而设计。从样例图像中提取颜色特征、形状特征和颜色区域的空间关系特征,并对其进行聚类,得到合适的取值范围。与只接受单个图像作为查询条件的QBE系统相比,MIERS (Multi-Image Example-based Retrieval System)的检索结果更好,实验结果表明,指定更多的样本有助于提高查全率,而查准率几乎没有下降。
{"title":"Content-based retrieval by multiple image examples for sign board retrieval","authors":"A. Yoshitaka, Terumasa Hyoudou","doi":"10.1109/PCS.2010.5702447","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702447","url":null,"abstract":"In the area of retrieving image databases, one of the promising approaches is to retrieve it by specifying image example. However, specifying a single image example is not always sufficient to get satisfactory result, since one image example does not give comprehensive ranges of values that reflect the various aspects of the object to be retrieved. In this paper, we propose a method of retrieving images by specifying multiple image examples that is designed for retrieving sign boards. Features of color, shape, and spatial relation of color regions are extracted from example images, and they are clustered so as to obtain proper range of values. Compared with QBE systems that accept only a single image as the query condition, MIERS (Multi-Image Example-based Retrieval System) returns better retrieval result, where the experimental result showed that specifying more examples helps to improve recall with little deterioration of precision.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114920988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A hierarchical variable-sized block transform coding scheme for coding efficiency improvement on H.264/AVC 一种提高H.264/AVC编码效率的分层变大小块变换编码方案
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702553
Bumshik Lee, Jae-il Kim, Sangsoo Ahn, Munchurl Kim, Hui-Yong Kim, Jong-Ho Kim, J. Choi
In this paper, a rate-distortion optimized variable block transform coding scheme is proposed based on a hierarchical structured transform for macroblock (MB) coding with a set of the order-4 and −8 integer cosine transform (ICT) kernels of H.264/AVC as well as a new order-16 ICT kernel. The set of order-4, −8 and −16 ICT kernels are applied for inter-predictive coding in square (4×4, 8×8 or 16×16) or non-square (16×8 or 8×16) transform for each MB in a hierarchical structured manner. The proposed hierarchical variable-sized block transform scheme using the order-16 ICT kernel achieves significant bitrate reduction up to 15%, compared to the High profile of H.264/AVC. Even if the number of candidates for the transform types increases, the encoding time can be reduced to average 4–6% over the H.264/AVC
本文利用H.264/AVC的4阶和- 8阶整数余弦变换核和16阶整数余弦变换核,提出了一种基于层次结构变换的率失真优化变块变换编码方案,用于宏块(MB)编码。将4阶、8阶和16阶ICT核集合以分层结构的方式对每个MB进行方形(4×4、8×8或16×16)或非方形(16×8或8×16)变换的预测编码。与H.264/AVC相比,采用16阶ICT内核的分层变大小块变换方案实现了高达15%的显着比特率降低。即使候选转换类型的数量增加,编码时间也可以比H.264/AVC平均减少4-6%
{"title":"A hierarchical variable-sized block transform coding scheme for coding efficiency improvement on H.264/AVC","authors":"Bumshik Lee, Jae-il Kim, Sangsoo Ahn, Munchurl Kim, Hui-Yong Kim, Jong-Ho Kim, J. Choi","doi":"10.1109/PCS.2010.5702553","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702553","url":null,"abstract":"In this paper, a rate-distortion optimized variable block transform coding scheme is proposed based on a hierarchical structured transform for macroblock (MB) coding with a set of the order-4 and −8 integer cosine transform (ICT) kernels of H.264/AVC as well as a new order-16 ICT kernel. The set of order-4, −8 and −16 ICT kernels are applied for inter-predictive coding in square (4×4, 8×8 or 16×16) or non-square (16×8 or 8×16) transform for each MB in a hierarchical structured manner. The proposed hierarchical variable-sized block transform scheme using the order-16 ICT kernel achieves significant bitrate reduction up to 15%, compared to the High profile of H.264/AVC. Even if the number of candidates for the transform types increases, the encoding time can be reduced to average 4–6% over the H.264/AVC","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114994715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Multiscale recurrent pattern matching approach for depth map coding 深度图编码的多尺度循环模式匹配方法
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702490
Danilo B. Graziosi, Nuno M. M. Rodrigues, C. Pagliari, E. Silva, S. Faria, Marcelo M. Perez, M. Carvalho
In this article we propose to compress depth maps using a coding scheme based on multiscale recurrent pattern matching and evaluate its impact on depth image based rendering (DIBR). Depth maps are usually converted into gray scale images and compressed like a conventional luminance signal. However, using traditional transform-based encoders to compress depth maps may result in undesired artifacts at sharp edges due to the quantization of high frequency coefficients. The Multidimensional Multiscale Parser (MMP) is a pattern matching-based encoder, that is able to preserve and efficiently encode high frequency patterns, such as edge information. This ability is critical for encoding depth map images. Experimental results for encoding depth maps show that MMP is much more efficient in a rate-distortion sense than standard image compression techniques such as JPEG2000 or H.264/AVC. In addition, the depth maps compressed with MMP generate reconstructed views with a higher quality than all other tested compression algorithms.
在本文中,我们提出了一种基于多尺度循环模式匹配的深度图压缩编码方案,并评估了其对基于深度图像的渲染(DIBR)的影响。深度图通常被转换成灰度图像,并像传统的亮度信号一样被压缩。然而,使用传统的基于变换的编码器来压缩深度图,由于高频系数的量化,可能会在尖锐边缘产生不希望的伪影。多维多尺度解析器(MMP)是一种基于模式匹配的编码器,能够有效地保存和编码高频模式,如边缘信息。这种能力对于编码深度图图像至关重要。深度图编码的实验结果表明,在率失真意义上,MMP比标准图像压缩技术(如JPEG2000或H.264/AVC)要高效得多。此外,与所有其他经过测试的压缩算法相比,用MMP压缩的深度图生成的重建视图质量更高。
{"title":"Multiscale recurrent pattern matching approach for depth map coding","authors":"Danilo B. Graziosi, Nuno M. M. Rodrigues, C. Pagliari, E. Silva, S. Faria, Marcelo M. Perez, M. Carvalho","doi":"10.1109/PCS.2010.5702490","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702490","url":null,"abstract":"In this article we propose to compress depth maps using a coding scheme based on multiscale recurrent pattern matching and evaluate its impact on depth image based rendering (DIBR). Depth maps are usually converted into gray scale images and compressed like a conventional luminance signal. However, using traditional transform-based encoders to compress depth maps may result in undesired artifacts at sharp edges due to the quantization of high frequency coefficients. The Multidimensional Multiscale Parser (MMP) is a pattern matching-based encoder, that is able to preserve and efficiently encode high frequency patterns, such as edge information. This ability is critical for encoding depth map images. Experimental results for encoding depth maps show that MMP is much more efficient in a rate-distortion sense than standard image compression techniques such as JPEG2000 or H.264/AVC. In addition, the depth maps compressed with MMP generate reconstructed views with a higher quality than all other tested compression algorithms.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116451860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
An epipolar resticted inter-mode selection for stereoscopic video encoding 一种用于立体视频编码的极限模式间选择
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702502
Guolei Yang, Luhong Liang, Wen Gao
Fast stereoscopic video encoding becomes a highly desired technique because the stereoscopic video has been realizable for applications like TV broadcasting and consumer electronics. The stereoscopic video has high inter-view dependency subject to epipolar restriction, which can be used to reduce the encoding complexity. In this paper, we propose a fast inter-prediction mode selection algorithm for stereoscopic video encoding. Different from methods using disparity estimation, candidate modes are generated by sliding a window along the macro-block line restricted by the epipolar. Then the motion information is utilized to rectify the candidate modes. A selection failure handling algorithm is also proposed to preserve coding quality. The proposed algorithm is evaluated using independent H.264/AVC encoders for left and right views and can be extended to MVC. Experimental results show that encoding times of one view are reduced by 41.4% and 24.4% for HD and VGA videos respectively with little quality loss.
快速立体视频编码成为一种非常需要的技术,因为立体视频已经在电视广播和消费电子产品等应用中实现。立体视频具有高度的视界间依赖性,受极域限制,可用于降低编码复杂度。本文提出了一种用于立体视频编码的快速预测模式选择算法。与使用视差估计的方法不同,候选模式是通过沿受极面限制的宏块线滑动窗口来产生的。然后利用运动信息对候选模式进行校正。为了保证编码质量,提出了一种选择失败处理算法。该算法使用独立的H.264/AVC编码器对左右视图进行了评估,并且可以扩展到MVC。实验结果表明,对于高清视频和VGA视频,单视图编码次数分别减少41.4%和24.4%,且质量损失较小。
{"title":"An epipolar resticted inter-mode selection for stereoscopic video encoding","authors":"Guolei Yang, Luhong Liang, Wen Gao","doi":"10.1109/PCS.2010.5702502","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702502","url":null,"abstract":"Fast stereoscopic video encoding becomes a highly desired technique because the stereoscopic video has been realizable for applications like TV broadcasting and consumer electronics. The stereoscopic video has high inter-view dependency subject to epipolar restriction, which can be used to reduce the encoding complexity. In this paper, we propose a fast inter-prediction mode selection algorithm for stereoscopic video encoding. Different from methods using disparity estimation, candidate modes are generated by sliding a window along the macro-block line restricted by the epipolar. Then the motion information is utilized to rectify the candidate modes. A selection failure handling algorithm is also proposed to preserve coding quality. The proposed algorithm is evaluated using independent H.264/AVC encoders for left and right views and can be extended to MVC. Experimental results show that encoding times of one view are reduced by 41.4% and 24.4% for HD and VGA videos respectively with little quality loss.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"64 245 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125959075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Two-dimensional Chebyshev polynomials for image fusion 二维切比雪夫多项式图像融合
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702526
Z. Omar, N. Mitianoudis, T. Stathaki
This report documents in detail the research carried out by the author throughout his first year. The paper presents a novel method for fusing images in a domain concerning multiple sensors and modalities. Using Chebyshev polynomials as basis functions, the image is decomposed to perform fusion at feature level. Results show favourable performance compared to previous efforts on image fusion, namely ICA and DT-CWT, in noise affected images. The work presented here aims at providing a novel framework for future studies in image analysis and may introduce innovations in the fields of surveillance, medical imaging and remote sensing.
这份报告详细记录了作者在第一年所进行的研究。提出了一种多传感器和多模态域图像融合的新方法。以切比雪夫多项式为基函数,对图像进行分解,在特征级进行融合。结果表明,在噪声影响的图像中,与之前的图像融合(即ICA和DT-CWT)相比,该方法具有良好的性能。本文介绍的工作旨在为今后的图像分析研究提供一个新的框架,并可能在监测、医学成像和遥感领域引入创新。
{"title":"Two-dimensional Chebyshev polynomials for image fusion","authors":"Z. Omar, N. Mitianoudis, T. Stathaki","doi":"10.1109/PCS.2010.5702526","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702526","url":null,"abstract":"This report documents in detail the research carried out by the author throughout his first year. The paper presents a novel method for fusing images in a domain concerning multiple sensors and modalities. Using Chebyshev polynomials as basis functions, the image is decomposed to perform fusion at feature level. Results show favourable performance compared to previous efforts on image fusion, namely ICA and DT-CWT, in noise affected images. The work presented here aims at providing a novel framework for future studies in image analysis and may introduce innovations in the fields of surveillance, medical imaging and remote sensing.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124745444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Avoidance of singular point in reversible KLT 可逆KLT中奇点的避免
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702435
M. Iwahashi, H. Kiya
In this report, permutation of order and sign of signals are introduced to avoid singular point problem of a reversible transform. When a transform is implemented in the lifting structure, it can be "reversible" in spite of rounding operations inside the transform. Therefore it has been applied to lossless coding of digital signals. However some coefficient values of the transform have singular points (SP). Around the SP, rounding errors are magnified to huge amount and the coding efficiency is decreased. In this report, we analyze the SP of a three point KLT for RGB color components of an image signal, and introduce permutation of order and sign of signals to avoid the SP problem. It was experimentally confirmed that the proposed method improved PSNR by approximately 15 [dB] comparing to the worst case.
为了避免可逆变换的奇点问题,本文引入了信号的序置换和符号置换。当在提升结构中实现转换时,它可以是“可逆的”,尽管在转换内部进行了舍入操作。因此,它已被应用于数字信号的无损编码。然而,变换的某些系数值存在奇异点。在SP附近,舍入误差被放大到很大,编码效率降低。在本报告中,我们分析了三点KLT对图像信号的RGB颜色分量的SP,并引入了信号的顺序和符号排列来避免SP问题。实验证实,与最坏情况相比,该方法的PSNR提高了约15 [dB]。
{"title":"Avoidance of singular point in reversible KLT","authors":"M. Iwahashi, H. Kiya","doi":"10.1109/PCS.2010.5702435","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702435","url":null,"abstract":"In this report, permutation of order and sign of signals are introduced to avoid singular point problem of a reversible transform. When a transform is implemented in the lifting structure, it can be \"reversible\" in spite of rounding operations inside the transform. Therefore it has been applied to lossless coding of digital signals. However some coefficient values of the transform have singular points (SP). Around the SP, rounding errors are magnified to huge amount and the coding efficiency is decreased. In this report, we analyze the SP of a three point KLT for RGB color components of an image signal, and introduce permutation of order and sign of signals to avoid the SP problem. It was experimentally confirmed that the proposed method improved PSNR by approximately 15 [dB] comparing to the worst case.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128537544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
28th Picture Coding Symposium
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1