首页 > 最新文献

28th Picture Coding Symposium最新文献

英文 中文
A subjective image quality metric for bit-inversion-based watermarking 基于位反转水印的主观图像质量度量
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702535
T. Kimoto, Fumihiko Kosaka
An image watermarking scheme using the previously proposed bit embedding method is developed. To achieve a desired subjective visual quality in the watermarked image, the embedding parameter that is related to both the image quality and the embedding capacity is determined by using a perceptual model. First, based on the properties of the bit embedding method, the perceptual model of two kinds of objective quality measures is assumed. Then, the measurements of human subjective image quality are analyzed from the viewpoint of the correlation with these two measures. Thereby, the estimating function that can yield an estimate of the subjective quality from two objective measurements is determined. According to the estimating function, the bit embedding method performs in each image region so as to achieve a desired subjective image quality while increasing the capacity of embedding watermark bits. The simulation results demonstrate that the estimating function values have a linear correlation with human subjective evaluations, and the embedding parameters can be adaptively changed in every image region by using the function.
提出了一种基于比特嵌入法的图像水印方案。为了在水印图像中获得理想的主观视觉质量,使用感知模型确定与图像质量和嵌入容量相关的嵌入参数。首先,基于位嵌入方法的特性,假设了两种客观质量度量的感知模型;然后,从两者的相关性角度分析了人类主观图像质量的测量方法。因此,确定了可以从两个客观测量中产生主观质量估计的估计函数。根据估计函数,在每个图像区域内进行位嵌入,在增加水印位嵌入容量的同时获得理想的主观图像质量。仿真结果表明,估计函数值与人的主观评价呈线性相关,并且利用该函数可以自适应地改变图像各区域的嵌入参数。
{"title":"A subjective image quality metric for bit-inversion-based watermarking","authors":"T. Kimoto, Fumihiko Kosaka","doi":"10.1109/PCS.2010.5702535","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702535","url":null,"abstract":"An image watermarking scheme using the previously proposed bit embedding method is developed. To achieve a desired subjective visual quality in the watermarked image, the embedding parameter that is related to both the image quality and the embedding capacity is determined by using a perceptual model. First, based on the properties of the bit embedding method, the perceptual model of two kinds of objective quality measures is assumed. Then, the measurements of human subjective image quality are analyzed from the viewpoint of the correlation with these two measures. Thereby, the estimating function that can yield an estimate of the subjective quality from two objective measurements is determined. According to the estimating function, the bit embedding method performs in each image region so as to achieve a desired subjective image quality while increasing the capacity of embedding watermark bits. The simulation results demonstrate that the estimating function values have a linear correlation with human subjective evaluations, and the embedding parameters can be adaptively changed in every image region by using the function.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114301558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An efficient side information generation using seed blocks for distributed video coding 基于种子块的分布式视频编码侧信息生成方法
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702585
Dong Yon Kim, Dongsan Jun, H. W. Park
Recently, a new video coding technique, distributed video coding (DVC), is an emerging research area for low power video coding applications. In the DVC, the encoder is much simpler than the conventional video codec, whereas the decoder is very heavy. The DVC decoder exploits side information which is generated by motion compensated frame interpolation to reconstruct the Wyner-Ziv frame. This paper proposes an efficient side information generation algorithm using seed blocks for DVC. Seed blocks are firstly selected to be used for motion estimation of the other blocks. As the side information is close to the target image, the final reconstructed image in the DVC decoder has better quality and the compression ratio becomes high. The proposed method contributes to improve the DVC compression performance with reduced computing time. Experimental results show that accurate motion vectors are estimated by the proposed method and its computational complexity of motion estimation is significantly reduced in comparison with the previous methods.
分布式视频编码(distributed video coding, DVC)是一种新的视频编码技术,是目前低功耗视频编码应用的一个新兴研究领域。在DVC中,编码器比传统的视频编解码器简单得多,而解码器却非常笨重。DVC解码器利用运动补偿帧插值产生的边信息重构Wyner-Ziv帧。提出了一种基于种子块的高效边信息生成算法。首先选择种子块用于其他块的运动估计。由于边线信息更接近目标图像,最终在DVC解码器中重构的图像质量更好,压缩比更高。该方法在提高DVC压缩性能的同时减少了计算时间。实验结果表明,该方法能够准确估计出运动矢量,且运动估计的计算量较以往方法显著降低。
{"title":"An efficient side information generation using seed blocks for distributed video coding","authors":"Dong Yon Kim, Dongsan Jun, H. W. Park","doi":"10.1109/PCS.2010.5702585","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702585","url":null,"abstract":"Recently, a new video coding technique, distributed video coding (DVC), is an emerging research area for low power video coding applications. In the DVC, the encoder is much simpler than the conventional video codec, whereas the decoder is very heavy. The DVC decoder exploits side information which is generated by motion compensated frame interpolation to reconstruct the Wyner-Ziv frame. This paper proposes an efficient side information generation algorithm using seed blocks for DVC. Seed blocks are firstly selected to be used for motion estimation of the other blocks. As the side information is close to the target image, the final reconstructed image in the DVC decoder has better quality and the compression ratio becomes high. The proposed method contributes to improve the DVC compression performance with reduced computing time. Experimental results show that accurate motion vectors are estimated by the proposed method and its computational complexity of motion estimation is significantly reduced in comparison with the previous methods.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128387021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Improving the visual quality of AVC/H.264 by combining it with content adaptive depth map compression 提高AVC/H的视觉质量。264与内容自适应深度图压缩相结合
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702545
Christian Keimel, K. Diepold, M. Sarkis
The future of video coding for 3DTV lies in the combination of depth maps and corresponding textures. Most current video coding standards, however, are only optimized for visual quality and are not able to efficiently compress depth maps. We present in this work a content adaptive depth map meshing with tritree and entropy encoding for 3D videos. We show that this approach outperforms the intra frame prediction of AVC/H.264 for the coding of depth maps of still images. We also demonstrate by combining AVC/H.264 with our algorithm that we are able to increase the visual quality of the encoded texture on average by 6 dB. This work is currently limited to still images but an extension to intra coding of 3D video is straightforward.
3DTV视频编码的未来在于深度图和相应纹理的结合。然而,目前大多数视频编码标准仅针对视觉质量进行了优化,而不能有效地压缩深度图。本文提出了一种基于三树和熵编码的3D视频内容自适应深度图网格划分方法。结果表明,该方法优于AVC/H的帧内预测。264用于对静止图像的深度图进行编码。我们还结合AVC/H进行了论证。使用我们的算法,我们能够将编码纹理的视觉质量平均提高6db。这项工作目前仅限于静止图像,但扩展到3D视频的内部编码是直截了当的。
{"title":"Improving the visual quality of AVC/H.264 by combining it with content adaptive depth map compression","authors":"Christian Keimel, K. Diepold, M. Sarkis","doi":"10.1109/PCS.2010.5702545","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702545","url":null,"abstract":"The future of video coding for 3DTV lies in the combination of depth maps and corresponding textures. Most current video coding standards, however, are only optimized for visual quality and are not able to efficiently compress depth maps. We present in this work a content adaptive depth map meshing with tritree and entropy encoding for 3D videos. We show that this approach outperforms the intra frame prediction of AVC/H.264 for the coding of depth maps of still images. We also demonstrate by combining AVC/H.264 with our algorithm that we are able to increase the visual quality of the encoded texture on average by 6 dB. This work is currently limited to still images but an extension to intra coding of 3D video is straightforward.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134618847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
H.264/AVC to wavelet-based scalable video transcoding supporting multiple coding configurations H.264/AVC到基于小波的可扩展视频转码,支持多种编码配置
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702564
Eduardo Peixoto, Toni Zgaljic, E. Izquierdo
Scalable Video Coding (SVC) enables low complexity adaptation of the compressed video, providing an efficient solution for video content delivery through heterogeneous networks and to different displays. However, legacy video and most commercially available content capturing devices use conventional non-scalable coding, e.g., H.264/AVC. This paper proposes an efficient transcoder from H.264/AVC to a wavelet-based SVC to exploit the advantages offerend by the SVC technology. The proposed transcoder is able to cope with different coding configurations in H.264/AVC, such as IPP or IBBP with multiple reference frames. To reduce the transcoder's complexity, motion information and presence of the residual data extracted from the decoded H.264/AVC video are exploited. Experimental results show a good performance of the proposed transcoder in terms of decoded video quality and system complexity.
可扩展视频编码(SVC)能够对压缩后的视频进行低复杂度的调整,为通过异构网络和不同显示器传输视频内容提供了一种有效的解决方案。然而,传统视频和大多数商业上可用的内容捕获设备使用传统的不可扩展编码,例如H.264/AVC。利用小波SVC技术的优势,提出了一种从H.264/AVC到基于小波SVC的高效转码器。所提出的转码器能够处理H.264/AVC中不同的编码配置,例如具有多个参考帧的IPP或IBBP。为了降低转码器的复杂度,利用了H.264/AVC视频中提取的运动信息和残差数据。实验结果表明,所提出的转码器在解码视频质量和系统复杂度方面具有良好的性能。
{"title":"H.264/AVC to wavelet-based scalable video transcoding supporting multiple coding configurations","authors":"Eduardo Peixoto, Toni Zgaljic, E. Izquierdo","doi":"10.1109/PCS.2010.5702564","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702564","url":null,"abstract":"Scalable Video Coding (SVC) enables low complexity adaptation of the compressed video, providing an efficient solution for video content delivery through heterogeneous networks and to different displays. However, legacy video and most commercially available content capturing devices use conventional non-scalable coding, e.g., H.264/AVC. This paper proposes an efficient transcoder from H.264/AVC to a wavelet-based SVC to exploit the advantages offerend by the SVC technology. The proposed transcoder is able to cope with different coding configurations in H.264/AVC, such as IPP or IBBP with multiple reference frames. To reduce the transcoder's complexity, motion information and presence of the residual data extracted from the decoded H.264/AVC video are exploited. Experimental results show a good performance of the proposed transcoder in terms of decoded video quality and system complexity.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"275 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134535569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
An adaptive early skip mode decision scheme for multiview video coding 一种多视点视频编码的自适应早期跳过模式决策方案
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702527
B. Zatt, M. Shafique, S. Bampi, J. Henkel
In this work a novel scheme is proposed for adaptive early SKIP mode decision in the multiview video coding based on mode correlation in the 3D-neighborhood, variance, and ratedistortion properties. Our scheme employs an adaptive thresholding mechanism in order to react to the changing values of Quantization Parameter (QP). Experimental results demonstrate that our scheme provides a consistent time saving over a wide range of QP values. Compared to the exhaustive mode decision, our scheme provides a significant reduction in the encoding complexity (up to 77%) at the cost of a small PSNR loss (0.172 dB in average). Compared to state-of-the-art, our scheme provides an average 2× higher complexity reduction with a relatively higher PSNR value (avg. 0.2 dB).
本文提出了一种基于3d邻域模式相关性、方差和率失真特性的多视点视频编码自适应早期SKIP模式决策的新方案。我们的方案采用自适应阈值机制,以应对量化参数(QP)值的变化。实验结果表明,我们的方案在大范围的QP值上提供了一致的时间节省。与穷举模式决策相比,我们的方案以较小的PSNR损失(平均0.172 dB)为代价,显著降低了编码复杂性(高达77%)。与最先进的方案相比,我们的方案提供了平均2倍的复杂性降低,具有相对较高的PSNR值(平均0.2 dB)。
{"title":"An adaptive early skip mode decision scheme for multiview video coding","authors":"B. Zatt, M. Shafique, S. Bampi, J. Henkel","doi":"10.1109/PCS.2010.5702527","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702527","url":null,"abstract":"In this work a novel scheme is proposed for adaptive early SKIP mode decision in the multiview video coding based on mode correlation in the 3D-neighborhood, variance, and ratedistortion properties. Our scheme employs an adaptive thresholding mechanism in order to react to the changing values of Quantization Parameter (QP). Experimental results demonstrate that our scheme provides a consistent time saving over a wide range of QP values. Compared to the exhaustive mode decision, our scheme provides a significant reduction in the encoding complexity (up to 77%) at the cost of a small PSNR loss (0.172 dB in average). Compared to state-of-the-art, our scheme provides an average 2× higher complexity reduction with a relatively higher PSNR value (avg. 0.2 dB).","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125967492","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Intra prediction architecture for H.264/AVC QFHD encoder H.264/AVC QFHD编码器的帧内预测结构
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702533
Gang He, Dajiang Zhou, Jinjia Zhou, S. Goto
This paper proposes a high-performance intra prediction architecture that can support H.264/AVC high profile. The proposed MB/block co-reordering can avoid data dependency and improve pipeline utilization. Therefore, the timing constraint of real-time 4k×2k encoding can be achieved with negligible quality loss. 16×16 prediction engine and 8×8 prediction engine work parallel for prediction and coefficients generating. A reordering interlaced reconstruction is also designed for fully pipelined architecture. It takes only 160 cycles to process one macroblock (MB). Hardware utilization of prediction and reconstruction modules is almost 100%. Furthermore, PE-reusable 8×8 intra predictor and hybrid SAD & SATD mode decision are proposed to save hardware cost. The design is implemented by 90nm CMOS technology with 113.2k gates and can encode 4k×2k video sequences at 60 fps with operation frequency of 310MHz.
本文提出了一种高性能的支持H.264/AVC的帧内预测体系结构。提出的MB/块协同重排序可以避免数据依赖,提高管道利用率。因此,实时4k×2k编码的时间约束可以在质量损失可以忽略不计的情况下实现。16×16预测引擎和8×8预测引擎并行工作,用于预测和系数生成。对于全流水线架构,还设计了一种重排序交错重构。处理一个宏块(MB)只需要160个周期。预测和重构模块的硬件利用率几乎为100%。此外,为了节省硬件成本,提出了pe可重用8×8内预测器和混合SAD和SATD模式决策。该设计采用90nm CMOS技术实现,采用113.2k栅极,以60fps的速度编码4k×2k视频序列,工作频率为310MHz。
{"title":"Intra prediction architecture for H.264/AVC QFHD encoder","authors":"Gang He, Dajiang Zhou, Jinjia Zhou, S. Goto","doi":"10.1109/PCS.2010.5702533","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702533","url":null,"abstract":"This paper proposes a high-performance intra prediction architecture that can support H.264/AVC high profile. The proposed MB/block co-reordering can avoid data dependency and improve pipeline utilization. Therefore, the timing constraint of real-time 4k×2k encoding can be achieved with negligible quality loss. 16×16 prediction engine and 8×8 prediction engine work parallel for prediction and coefficients generating. A reordering interlaced reconstruction is also designed for fully pipelined architecture. It takes only 160 cycles to process one macroblock (MB). Hardware utilization of prediction and reconstruction modules is almost 100%. Furthermore, PE-reusable 8×8 intra predictor and hybrid SAD & SATD mode decision are proposed to save hardware cost. The design is implemented by 90nm CMOS technology with 113.2k gates and can encode 4k×2k video sequences at 60 fps with operation frequency of 310MHz.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124915004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Temporal inconsistency measure for video quality assessment 视频质量评价的时间不一致性度量
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702572
Songnan Li, Lin Ma, Fan Zhang, K. Ngan
Visual quality assessment plays a crucial role in many vision-related signal processing applications. In the literature, more efforts have been spent on spatial visual quality measure. Although a large number of video quality metrics have been proposed, the methods to use temporal information for quality assessment are less diversified. In this paper, we propose a novel method to measure the temporal impairments. The proposed method can be incorporated into any image quality metric to extend it into a video quality metric. Moreover, it is easy to apply the proposed method in video coding system to incorporate with MSE for rate-distortion optimization.
视觉质量评估在许多与视觉相关的信号处理应用中起着至关重要的作用。在文献中,对空间视觉质量测量的研究较多。尽管已经提出了大量的视频质量度量,但使用时间信息进行质量评估的方法并不多样化。本文提出了一种测量颞叶损伤的新方法。所提出的方法可以结合到任何图像质量度量中,将其扩展为视频质量度量。此外,该方法易于应用于视频编码系统中,结合MSE进行码率失真优化。
{"title":"Temporal inconsistency measure for video quality assessment","authors":"Songnan Li, Lin Ma, Fan Zhang, K. Ngan","doi":"10.1109/PCS.2010.5702572","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702572","url":null,"abstract":"Visual quality assessment plays a crucial role in many vision-related signal processing applications. In the literature, more efforts have been spent on spatial visual quality measure. Although a large number of video quality metrics have been proposed, the methods to use temporal information for quality assessment are less diversified. In this paper, we propose a novel method to measure the temporal impairments. The proposed method can be incorporated into any image quality metric to extend it into a video quality metric. Moreover, it is easy to apply the proposed method in video coding system to incorporate with MSE for rate-distortion optimization.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"84 11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126023333","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Compressed signature for video identification 压缩签名视频识别
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702534
N. Sprljan, P. Brasnett, S. Paschalakis
This paper presents a new application-specific lossless compression scheme developed for video identification descriptors, also known as video fingerprints or signatures. In designing such a descriptor, one usually has to balance the descriptor size against discriminating power and temporal localisation performance. The proposed compression scheme alleviates this problem by efficiently exploiting the temporal redundancies present in the video fingerprint, allowing highly accurate fingerprints which also entail low transmission and storage costs. In this paper we provide a detailed description of our compression scheme and a comparative evaluation against well known state-of-the-art generic compression tools.
本文提出了一种针对视频识别描述符(也称为视频指纹或视频签名)的专用无损压缩方案。在设计这样的描述符时,通常需要平衡描述符的大小、判别能力和时间定位性能。所提出的压缩方案通过有效地利用视频指纹中存在的时间冗余来缓解这一问题,从而实现高精度指纹,同时降低传输和存储成本。在本文中,我们提供了一个详细的描述,我们的压缩方案和比较评估,对众所周知的最先进的通用压缩工具。
{"title":"Compressed signature for video identification","authors":"N. Sprljan, P. Brasnett, S. Paschalakis","doi":"10.1109/PCS.2010.5702534","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702534","url":null,"abstract":"This paper presents a new application-specific lossless compression scheme developed for video identification descriptors, also known as video fingerprints or signatures. In designing such a descriptor, one usually has to balance the descriptor size against discriminating power and temporal localisation performance. The proposed compression scheme alleviates this problem by efficiently exploiting the temporal redundancies present in the video fingerprint, allowing highly accurate fingerprints which also entail low transmission and storage costs. In this paper we provide a detailed description of our compression scheme and a comparative evaluation against well known state-of-the-art generic compression tools.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130113019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Challenges in multiview video — The 3 D'S 多视点视频的挑战- 3d
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702454
Antonio Ortega
For the purpose of this paper we group, under the generic term multiview video, different systems for which multiple standard video cameras and, possibly, additional depth-capturing cameras, are used. Video is then presented to the user using special glasses or displays. Research work in this area has focused on topics ranging from designing compression techniques to developing new 3D displays. In this paper we primarily consider the challenges involved in developing efficient compression tools. Our primary observation is that the “right” coding tools could depend heavily on choices made for content capture, display and communication. This is of course true for conventional video coding as well. But we will argue that it is even more important to address these issues for multiview video because there are much greater differences between different application scenarios (as compared to conventional video). The risk is that coding tools that are too narrowly focused on a specific application scenario may not be at all suitable for others. We focus specifically on three factors for which there exists significant uncertainty, namely, displays, depth estimation and content delivery. Our goal is not to discuss in detail current and future approaches (e.g., emerging alternative display technologies), but rather to show how these various approaches may have an impact on compression system design.
为了本文的目的,我们在通用术语“多视点视频”下,对使用多个标准视频摄像机和可能使用的额外深度捕捉摄像机的不同系统进行分组。然后,视频通过特殊的眼镜或显示器呈现给用户。该领域的研究工作集中在从设计压缩技术到开发新的3D显示器等主题上。在本文中,我们主要考虑开发高效压缩工具所涉及的挑战。我们的主要观察是,“正确的”编码工具可能在很大程度上取决于为内容捕获、显示和交流所做的选择。当然,对于传统的视频编码也是如此。但我们认为,解决这些问题对于多视图视频更为重要,因为不同应用场景之间存在更大的差异(与传统视频相比)。风险在于过于狭隘地专注于特定应用程序场景的编码工具可能根本不适合其他应用程序。我们特别关注存在显著不确定性的三个因素,即显示、深度估计和内容交付。我们的目标不是详细讨论当前和未来的方法(例如,新兴的替代显示技术),而是展示这些不同的方法如何对压缩系统设计产生影响。
{"title":"Challenges in multiview video — The 3 D'S","authors":"Antonio Ortega","doi":"10.1109/PCS.2010.5702454","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702454","url":null,"abstract":"For the purpose of this paper we group, under the generic term multiview video, different systems for which multiple standard video cameras and, possibly, additional depth-capturing cameras, are used. Video is then presented to the user using special glasses or displays. Research work in this area has focused on topics ranging from designing compression techniques to developing new 3D displays. In this paper we primarily consider the challenges involved in developing efficient compression tools. Our primary observation is that the “right” coding tools could depend heavily on choices made for content capture, display and communication. This is of course true for conventional video coding as well. But we will argue that it is even more important to address these issues for multiview video because there are much greater differences between different application scenarios (as compared to conventional video). The risk is that coding tools that are too narrowly focused on a specific application scenario may not be at all suitable for others. We focus specifically on three factors for which there exists significant uncertainty, namely, displays, depth estimation and content delivery. Our goal is not to discuss in detail current and future approaches (e.g., emerging alternative display technologies), but rather to show how these various approaches may have an impact on compression system design.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126306111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D pose estimation in high dimensional search spaces with local memorization 基于局部记忆的高维搜索空间三维姿态估计
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702507
Weilan Luo, T. Yamasaki, K. Aizawa
In this paper, a stochastic approach for extracting the articulated 3D human postures by synchronized multiple cameras in the high-dimensional configuration spaces is presented. Annealed Particle Filtering (APF) [1] seeks for the globally optimal solution of the likelihood. We improve and extend the APF with local memorization to estimate the suited kinematic postures for a volume sequence directly instead of projecting a rough simplified body model to 2D images. Our method guides the particles to the global optimization on the basis of local constraints. A segmentation algorithm is performed on the volumetric models and the process is repeated. We assign the articulated models 42 degrees of freedom. The matching error is about 6% on average while tracking the posture between two neighboring frames.
本文提出了一种基于同步多摄像机的高维人体姿态随机提取方法。退火粒子滤波(APF)[1]寻求似然的全局最优解。我们改进和扩展了局部记忆的APF,直接估计适合体序列的运动学姿态,而不是将粗糙的简化身体模型投影到二维图像上。我们的方法在局部约束的基础上引导粒子进行全局优化。在体积模型上执行分割算法,并重复该过程。我们赋予铰接模型42个自由度。在两个相邻帧之间跟踪姿态时,匹配误差平均约为6%。
{"title":"3D pose estimation in high dimensional search spaces with local memorization","authors":"Weilan Luo, T. Yamasaki, K. Aizawa","doi":"10.1109/PCS.2010.5702507","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702507","url":null,"abstract":"In this paper, a stochastic approach for extracting the articulated 3D human postures by synchronized multiple cameras in the high-dimensional configuration spaces is presented. Annealed Particle Filtering (APF) [1] seeks for the globally optimal solution of the likelihood. We improve and extend the APF with local memorization to estimate the suited kinematic postures for a volume sequence directly instead of projecting a rough simplified body model to 2D images. Our method guides the particles to the global optimization on the basis of local constraints. A segmentation algorithm is performed on the volumetric models and the process is repeated. We assign the articulated models 42 degrees of freedom. The matching error is about 6% on average while tracking the posture between two neighboring frames.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126439158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
28th Picture Coding Symposium
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1