首页 > 最新文献

2010 IEEE International Workshop on Multimedia Signal Processing最新文献

英文 中文
Depth-aided image inpainting for novel view synthesis 用于新型视图合成的深度辅助图像绘制
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5662013
Ismaël Daribo, B. Pesquet-Popescu
Depth Image Based Rendering (DIBR) technique has been recognized as a promising tool for supporting advanced 3D video services required in MultiView Video (MVV) systems. However, an inherent problem with DIBR is to fill newly exposed areas (holes) caused by disocclusions. This paper addresses the disocclusion problem. To deal with small disocclusions, hole-filling strategies have been designed by the state-of-the-art through pre-processing techniques of the depth video. For larger disocclusions, where depth pre-processing has some limitations, we propose an inpainting approach to retrieve missing pixels. Specifically, we propose in the texture and structure propagation process to take into account the depth information by distinguishing foreground and background parts of the scene. Experimental results illustrate the efficiency of the proposed method.
深度图像渲染(DIBR)技术被认为是支持多视图视频(MVV)系统所需的高级3D视频服务的一种有前途的工具。然而,DIBR的一个固有问题是填充由咬合引起的新暴露区域(孔)。本文讨论了咬合问题。为了处理小的咬合,通过深度视频的预处理技术,设计了最先进的补孔策略。对于较大的咬合,深度预处理有一定的局限性,我们提出了一种修复方法来检索缺失的像素。具体来说,我们提出在纹理和结构传播过程中,通过区分场景的前景和背景部分来考虑深度信息。实验结果表明了该方法的有效性。
{"title":"Depth-aided image inpainting for novel view synthesis","authors":"Ismaël Daribo, B. Pesquet-Popescu","doi":"10.1109/MMSP.2010.5662013","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5662013","url":null,"abstract":"Depth Image Based Rendering (DIBR) technique has been recognized as a promising tool for supporting advanced 3D video services required in MultiView Video (MVV) systems. However, an inherent problem with DIBR is to fill newly exposed areas (holes) caused by disocclusions. This paper addresses the disocclusion problem. To deal with small disocclusions, hole-filling strategies have been designed by the state-of-the-art through pre-processing techniques of the depth video. For larger disocclusions, where depth pre-processing has some limitations, we propose an inpainting approach to retrieve missing pixels. Specifically, we propose in the texture and structure propagation process to take into account the depth information by distinguishing foreground and background parts of the scene. Experimental results illustrate the efficiency of the proposed method.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127185437","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 151
Multistage compressed-sensing reconstruction of multiview images 多视角图像的多级压缩感知重构
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5662003
M. Trocan, Thomas Maugey, Eric W. Tramel, J. Fowler, B. Pesquet-Popescu
Compressed sensing is applied to multiview image sets and the high degree of correlation between views is exploited to enhance recovery performance over straightforward independent view recovery. This gain in performance is obtained by recovering the difference between a set of acquired measurements and the projection of a prediction of the signal they represent. The recovered difference is then added back to the prediction, and the prediction and recovery procedure is repeated in an iterated fashion for each of the views in the multiview image set. The recovered multiview image set is then used as an initialization to repeat the entire process again to form a multistage refinement. Experimental results reveal substantial performance gains from the multistage reconstruction.
将压缩感知应用于多视图图像集,利用视图之间的高度相关性来提高恢复性能,而不是直接的独立视图恢复。这种性能增益是通过恢复一组采集的测量值与它们所代表的信号的预测投影之间的差来获得的。然后将恢复的差异添加回预测中,并以迭代的方式对多视图图像集中的每个视图重复预测和恢复过程。然后将恢复的多视图图像集用作初始化,再次重复整个过程以形成多阶段细化。实验结果表明,多级重构可显著提高性能。
{"title":"Multistage compressed-sensing reconstruction of multiview images","authors":"M. Trocan, Thomas Maugey, Eric W. Tramel, J. Fowler, B. Pesquet-Popescu","doi":"10.1109/MMSP.2010.5662003","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5662003","url":null,"abstract":"Compressed sensing is applied to multiview image sets and the high degree of correlation between views is exploited to enhance recovery performance over straightforward independent view recovery. This gain in performance is obtained by recovering the difference between a set of acquired measurements and the projection of a prediction of the signal they represent. The recovered difference is then added back to the prediction, and the prediction and recovery procedure is repeated in an iterated fashion for each of the views in the multiview image set. The recovered multiview image set is then used as an initialization to repeat the entire process again to form a multistage refinement. Experimental results reveal substantial performance gains from the multistage reconstruction.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131688202","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Robust foreground segmentation for GPU architecture in an immersive 3D videoconferencing system 沉浸式3D视频会议系统中GPU架构的鲁棒前景分割
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5661997
J. Civit, Ò. Escoda
Current telepresence systems, while being a great step forward in videoconferencing, still have important points to improve in what eye-contact, gaze and gesture awareness concerns. Many-to-many communications are going to greatly benefit from mature auto-stereoscopic 3D technology; allowing people to engage more natural remote meetings, with proper eye-contact and better spatiality feeling. For this purpose, proper real-time multi-perspective 3D video capture is necessary (often based on one or more View+Depth data sets). Given current state of the art, some sort of foreground segmentation is often necessary at the acquisition in order to generate 3D depth maps with hight enough resolution and accurate object boundaries. For this, one needs flicker-less foreground segmentations, accurate to borders, resilient to noise and foreground shade changes, and able to operate in real-time on performing architectures such as GPGPUs. This paper introduces a robust Foreground Segmentation approach used within the experimental immersive 3D Telepresence system from EU-FP7 3DPresence project. The proposed algorithm is based on a costs minimization using Hierarchical Believe Propagation and outliers reduction by regularization on oversegmented regions. The iterative nature of the approach makes it scalable in complexity, allowing it to increase accuracy and picture size capacity as GPGPUs become faster. In this work, particular care in the design of foreground and background cost models has also been taken in order to overcome limitations of previous work proposed in the literature.
目前的远程呈现系统,虽然是视频会议的一大进步,但在目光接触、凝视和手势感知方面仍有重要的改进之处。多对多通信将从成熟的自动立体3D技术中受益匪浅;让人们参与更自然的远程会议,适当的眼神交流和更好的空间感。为此,适当的实时多视角3D视频捕获是必要的(通常基于一个或多个View+Depth数据集)。考虑到目前的技术水平,为了生成具有足够高分辨率和精确目标边界的3D深度图,在采集过程中通常需要某种前景分割。为此,人们需要无闪烁的前景分割,精确的边界,对噪声和前景阴影变化的弹性,并能够在诸如gpgpu之类的执行架构上实时操作。本文介绍了一种在EU-FP7 3DPresence项目的实验性沉浸式3D远程呈现系统中使用的鲁棒前景分割方法。该算法基于分层置信传播的成本最小化和对过度分割区域的正则化异常值减少。该方法的迭代特性使其在复杂性上具有可扩展性,允许它在gpgpu变得更快时提高准确性和图像大小容量。在这项工作中,为了克服文献中提出的先前工作的局限性,还特别注意了前景和背景成本模型的设计。
{"title":"Robust foreground segmentation for GPU architecture in an immersive 3D videoconferencing system","authors":"J. Civit, Ò. Escoda","doi":"10.1109/MMSP.2010.5661997","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5661997","url":null,"abstract":"Current telepresence systems, while being a great step forward in videoconferencing, still have important points to improve in what eye-contact, gaze and gesture awareness concerns. Many-to-many communications are going to greatly benefit from mature auto-stereoscopic 3D technology; allowing people to engage more natural remote meetings, with proper eye-contact and better spatiality feeling. For this purpose, proper real-time multi-perspective 3D video capture is necessary (often based on one or more View+Depth data sets). Given current state of the art, some sort of foreground segmentation is often necessary at the acquisition in order to generate 3D depth maps with hight enough resolution and accurate object boundaries. For this, one needs flicker-less foreground segmentations, accurate to borders, resilient to noise and foreground shade changes, and able to operate in real-time on performing architectures such as GPGPUs. This paper introduces a robust Foreground Segmentation approach used within the experimental immersive 3D Telepresence system from EU-FP7 3DPresence project. The proposed algorithm is based on a costs minimization using Hierarchical Believe Propagation and outliers reduction by regularization on oversegmented regions. The iterative nature of the approach makes it scalable in complexity, allowing it to increase accuracy and picture size capacity as GPGPUs become faster. In this work, particular care in the design of foreground and background cost models has also been taken in order to overcome limitations of previous work proposed in the literature.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"43 10","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131751620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Depth consistency testing for improved view interpolation 深度一致性测试改进的视图插值
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5662051
P. Rana, M. Flierl
Multiview video will play a pivotal role in the next generation visual communication media services like three-dimensional (3D) television and free-viewpoint television. These advanced media services provide natural 3D impressions and enable viewers to move freely in a dynamic real world scene by changing the viewpoint. High quality virtual view interpolation is required to support free viewpoint viewing. Usually, depth maps of different viewpoints are used to reconstruct a novel view. As these depth maps are usually estimated individually by stereo-matching algorithms, they have very weak spatial consistency. The inconsistency of depth maps affects the quality of view interpolation. In this paper, we propose a method for depth consistency testing to improve view interpolation. The method addresses the problem by warping more than two depth maps from multiple reference viewpoints to the virtual viewpoint. We test the consistency among warped depth values and improve the depth value information of the virtual view. With that, we enhance the quality of the interpolated virtual view.
多视点视频将在三维电视和自由视点电视等下一代视觉传播媒体服务中发挥关键作用。这些先进的媒体服务提供自然的3D印象,使观众能够通过改变视点在动态的现实世界场景中自由移动。高质量的虚拟视图插值需要支持自由视点观看。通常使用不同视点的深度图来重建一个新的视图。由于这些深度图通常是通过立体匹配算法单独估计的,因此它们的空间一致性非常弱。深度图的不一致性影响了视图插值的质量。在本文中,我们提出了一种深度一致性测试方法来改进视图插值。该方法通过将多个参考视点的两个以上深度图翘曲到虚拟视点来解决这个问题。我们测试了扭曲深度值之间的一致性,并改进了虚拟视图的深度值信息。这样,我们就提高了插值虚拟视图的质量。
{"title":"Depth consistency testing for improved view interpolation","authors":"P. Rana, M. Flierl","doi":"10.1109/MMSP.2010.5662051","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5662051","url":null,"abstract":"Multiview video will play a pivotal role in the next generation visual communication media services like three-dimensional (3D) television and free-viewpoint television. These advanced media services provide natural 3D impressions and enable viewers to move freely in a dynamic real world scene by changing the viewpoint. High quality virtual view interpolation is required to support free viewpoint viewing. Usually, depth maps of different viewpoints are used to reconstruct a novel view. As these depth maps are usually estimated individually by stereo-matching algorithms, they have very weak spatial consistency. The inconsistency of depth maps affects the quality of view interpolation. In this paper, we propose a method for depth consistency testing to improve view interpolation. The method addresses the problem by warping more than two depth maps from multiple reference viewpoints to the virtual viewpoint. We test the consistency among warped depth values and improve the depth value information of the virtual view. With that, we enhance the quality of the interpolated virtual view.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132102870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Optimizing the free distance of Error-Correcting Variable-Length Codes 纠错变长码自由距离的优化
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5662027
A. Diallo, C. Weidmann, M. Kieffer
This paper considers the optimization of Error-Correcting Variable-Length Codes (EC-VLC), which are a class of joint-source channel codes. The aim is to find a prefix-free codebook with the largest possible free distance for a given set of codeword lengths, ℓ = (ℓ1, ℓ2, …, ℓM). The proposed approach consists in ordering all possible codebooks associated to ℓ on a tree, and then to apply an efficient branch-and-prune algorithm to find a codebook with maximal free distance. Three methods for building the tree of codebooks are presented and their efficiency is compared.
本文研究了一类联合源信道码——纠错变长码(EC-VLC)的优化问题。目标是找到一个无前缀的码本,对于给定的码字长度集,具有最大可能的自由距离,为(1,2,…,M)。提出的方法是对树上所有可能的与l相关的码本进行排序,然后应用一个有效的分支-剪枝算法来寻找自由距离最大的码本。提出了三种构建码本树的方法,并对其效率进行了比较。
{"title":"Optimizing the free distance of Error-Correcting Variable-Length Codes","authors":"A. Diallo, C. Weidmann, M. Kieffer","doi":"10.1109/MMSP.2010.5662027","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5662027","url":null,"abstract":"This paper considers the optimization of Error-Correcting Variable-Length Codes (EC-VLC), which are a class of joint-source channel codes. The aim is to find a prefix-free codebook with the largest possible free distance for a given set of codeword lengths, ℓ = (ℓ<inf>1</inf>, ℓ<inf>2</inf>, …, ℓ<inf>M</inf>). The proposed approach consists in ordering all possible codebooks associated to ℓ on a tree, and then to apply an efficient branch-and-prune algorithm to find a codebook with maximal free distance. Three methods for building the tree of codebooks are presented and their efficiency is compared.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129970313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Scalable-to-lossless transform domain distributed video coding 可缩放到无损变换域分布式视频编码
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5662041
Xin Huang, Anna Ukhanova, A. Veselov, Søren Forchhammer, M. Gilmutdinov
Distributed video coding (DVC) is a novel approach providing new features as low complexity encoding by mainly exploiting the source statistics at the decoder based on the availability of decoder side information. In this paper, scalable-to-lossless DVC is presented based on extending a lossy Transform Domain Wyner-Ziv (TDWZ) distributed video codec with feedback. The lossless coding is obtained by using a reversible integer DCT. Experimental results show that the performance of the proposed scalable-to-lossless TDWZ video codec can outperform alternatives based on the JPEG 2000 standard. The TDWZ codec provides frame by frame encoding. Comparing the lossless coding efficiency, the proposed scalable-to-lossless TDWZ video codec can save up to 5%-13% bits compared to JPEG LS and H.264 Intra frame lossless coding and do so as a scalable-to-lossless coding.
分布式视频编码(DVC)是一种新颖的编码方法,它主要利用解码器侧信息的可用性来利用解码器端的源统计信息,从而提供了低复杂度编码的新特性。本文在对有损变换域Wyner-Ziv (TDWZ)分布式视频编解码器进行反馈扩展的基础上,提出了可扩展到无损DVC的方法。采用可逆整数DCT进行无损编码。实验结果表明,所提出的可扩展至无损的TDWZ视频编解码器的性能优于基于JPEG 2000标准的备选方案。TDWZ编解码器提供逐帧编码。通过对无损编码效率的比较,与JPEG LS和H.264的帧内无损编码相比,所提出的可扩展到无损的TDWZ视频编解码器可节省5% ~ 13%的比特,是一种可扩展到无损编码。
{"title":"Scalable-to-lossless transform domain distributed video coding","authors":"Xin Huang, Anna Ukhanova, A. Veselov, Søren Forchhammer, M. Gilmutdinov","doi":"10.1109/MMSP.2010.5662041","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5662041","url":null,"abstract":"Distributed video coding (DVC) is a novel approach providing new features as low complexity encoding by mainly exploiting the source statistics at the decoder based on the availability of decoder side information. In this paper, scalable-to-lossless DVC is presented based on extending a lossy Transform Domain Wyner-Ziv (TDWZ) distributed video codec with feedback. The lossless coding is obtained by using a reversible integer DCT. Experimental results show that the performance of the proposed scalable-to-lossless TDWZ video codec can outperform alternatives based on the JPEG 2000 standard. The TDWZ codec provides frame by frame encoding. Comparing the lossless coding efficiency, the proposed scalable-to-lossless TDWZ video codec can save up to 5%-13% bits compared to JPEG LS and H.264 Intra frame lossless coding and do so as a scalable-to-lossless coding.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"160 9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129024354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Error concealment considering error propagation inside a frame 考虑帧内错误传播的错误隐藏
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5662053
Jun Wang, Yichun Tang, Hao Sun, S. Goto
Transmission of compressed video over error prone channels may result in packet losses or errors, which can significantly degrade the image quality. Such degradation even becomes worse in 1Seg video broadcasting, which is widely used in Japan and Brazil for mobile phone TV service recently, where errors are drastically increased and lost areas are contiguous. Therefore the errors in earlier concealed MBs (macro blocks) may propagate to the MBs later to be concealed inside the same frame (spatial domain). The error concealment (EC) is used to recover the lost data by the redundancy in videos. Aiming at spatial error propagation (SEP) reduction, this paper proposes a SEP reduction based EC (SEPEC). In SEPEC, besides the mismatch distortion in current MB, the potential propagated mismatch distortion in the following to be concealed MBs is also minimized. Also, 2 extensions of SEPEC, that SEPEC with refined search and SEPEC with multiple layer match are discussed. Compared with previous work, the experiments show SEPEC achieves much better performance of video recovery and excellent trade-off between quality and computation in 1Seg broadcasting in terms of computation cost.
在容易出错的信道上传输压缩视频可能会导致丢包或错误,从而严重降低图像质量。最近在日本和巴西的手机电视服务中广泛使用的1Seg视频广播中,错误急剧增加,而且丢失区域是连续的,这种退化现象更加严重。因此,先前隐藏的mb(宏块)中的错误可能会传播到稍后隐藏在同一帧(空间域)中的mb中。错误隐藏(EC)是利用视频中的冗余来恢复丢失的数据。针对空间误差传播(SEP)抑制问题,提出了一种基于SEP抑制的EC (SEPEC)算法。在SEPEC中,除了当前MB中的失配失真外,后续MB中潜在的传播失配失真也被最小化。讨论了SEPEC的两种扩展,即带精细化搜索的SEPEC和带多层匹配的SEPEC。实验结果表明,与以往的工作相比,SEPEC在1Seg广播中获得了更好的视频恢复性能,并且在计算成本方面实现了良好的质量与计算之间的平衡。
{"title":"Error concealment considering error propagation inside a frame","authors":"Jun Wang, Yichun Tang, Hao Sun, S. Goto","doi":"10.1109/MMSP.2010.5662053","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5662053","url":null,"abstract":"Transmission of compressed video over error prone channels may result in packet losses or errors, which can significantly degrade the image quality. Such degradation even becomes worse in 1Seg video broadcasting, which is widely used in Japan and Brazil for mobile phone TV service recently, where errors are drastically increased and lost areas are contiguous. Therefore the errors in earlier concealed MBs (macro blocks) may propagate to the MBs later to be concealed inside the same frame (spatial domain). The error concealment (EC) is used to recover the lost data by the redundancy in videos. Aiming at spatial error propagation (SEP) reduction, this paper proposes a SEP reduction based EC (SEPEC). In SEPEC, besides the mismatch distortion in current MB, the potential propagated mismatch distortion in the following to be concealed MBs is also minimized. Also, 2 extensions of SEPEC, that SEPEC with refined search and SEPEC with multiple layer match are discussed. Compared with previous work, the experiments show SEPEC achieves much better performance of video recovery and excellent trade-off between quality and computation in 1Seg broadcasting in terms of computation cost.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"140 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117040068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Optimized decomposition basis using Lanczos filters for lossless compression of biomedical images 利用Lanczos滤波器优化分解基础,对生物医学图像进行无损压缩
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5662005
Jonathan Taquet, C. Labit
This paper proposes to introduce Lanczos interpolation filters as wavelet atoms in an optimized decomposition for embedded lossy to lossless compression of biomedical images. The decomposition and the Lanczos parameter are jointly optimized in a generic packet structure in order to take into account the various contents of biomedical imaging modalities. Lossless experimental results are given on a large scale database. They show that in comparison with a well known basis using 5/3 biorthogonal wavelets and a dyadic decomposition, the proposed approach allows to improve the compression by more than 10% on less noisy images and up to 30% on 3D-MRI while providing similar results on noisy datasets.
本文提出将Lanczos插值滤波器作为小波原子引入生物医学图像嵌入有损压缩到无损压缩的优化分解中。为了考虑生物医学成像模态的多种内容,在通用包结构中对分解和Lanczos参数进行了联合优化。在大型数据库上给出了无损实验结果。他们表明,与使用5/3双正交小波和二进分解的已知基础相比,所提出的方法可以在低噪声图像上提高10%以上的压缩率,在3D-MRI上提高30%,同时在噪声数据集上提供类似的结果。
{"title":"Optimized decomposition basis using Lanczos filters for lossless compression of biomedical images","authors":"Jonathan Taquet, C. Labit","doi":"10.1109/MMSP.2010.5662005","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5662005","url":null,"abstract":"This paper proposes to introduce Lanczos interpolation filters as wavelet atoms in an optimized decomposition for embedded lossy to lossless compression of biomedical images. The decomposition and the Lanczos parameter are jointly optimized in a generic packet structure in order to take into account the various contents of biomedical imaging modalities. Lossless experimental results are given on a large scale database. They show that in comparison with a well known basis using 5/3 biorthogonal wavelets and a dyadic decomposition, the proposed approach allows to improve the compression by more than 10% on less noisy images and up to 30% on 3D-MRI while providing similar results on noisy datasets.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122568373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Motion vector forecast and mapping (MV-FMap) method for entropy coding based video coders 基于熵编码的视频编码器的运动矢量预测和映射(MV-FMap)方法
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5662020
J. L. Tanou, Jean-Marc Thiesse, Joël Jung, M. Antonini
Since the finalization of the H.264/AVC standard and in order to meet the target set by both ITU-T and MPEG to define a new standard that reaches 50% bit rate reduction compared to H.264/AVC, many tools have efficiently improved the texture coding and the motion compensation accuracy. These improvements have resulted in increasing the proportion of bit rate allocated to motion information. Thus, the bit rate reduction of this information becomes a key subject of research. This paper proposes a method for motion vector coding based on an adaptive redistribution of motion vector residuals before entropy coding. Motion information is gathered to forecast a list of motion vector residuals which are redistributed to unexpected residuals of lower coding cost. Compared to H.264/AVC, this scheme provides systematic gain on tested sequences, and 2.3% in average, reaching up to 4.9% for a given sequence.
自H.264/AVC标准最终确定以来,为了满足ITU-T和MPEG共同设定的目标,即定义一个比H.264/AVC降低50%比特率的新标准,许多工具都有效地提高了纹理编码和运动补偿精度。这些改进增加了分配给运动信息的比特率比例。因此,降低这些信息的比特率成为研究的重点。提出了一种基于熵编码前运动矢量残差自适应重分布的运动矢量编码方法。收集运动信息来预测运动矢量残差列表,这些残差被重新分配到编码成本较低的意外残差中。与H.264/AVC相比,该方案在测试序列上提供系统增益,平均增益为2.3%,对于给定序列可达4.9%。
{"title":"Motion vector forecast and mapping (MV-FMap) method for entropy coding based video coders","authors":"J. L. Tanou, Jean-Marc Thiesse, Joël Jung, M. Antonini","doi":"10.1109/MMSP.2010.5662020","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5662020","url":null,"abstract":"Since the finalization of the H.264/AVC standard and in order to meet the target set by both ITU-T and MPEG to define a new standard that reaches 50% bit rate reduction compared to H.264/AVC, many tools have efficiently improved the texture coding and the motion compensation accuracy. These improvements have resulted in increasing the proportion of bit rate allocated to motion information. Thus, the bit rate reduction of this information becomes a key subject of research. This paper proposes a method for motion vector coding based on an adaptive redistribution of motion vector residuals before entropy coding. Motion information is gathered to forecast a list of motion vector residuals which are redistributed to unexpected residuals of lower coding cost. Compared to H.264/AVC, this scheme provides systematic gain on tested sequences, and 2.3% in average, reaching up to 4.9% for a given sequence.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122810125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A comparative study between different pre-whitening decorrelation based acoustic feedback cancellers 基于预白化去相关的不同声反馈消除器的比较研究
Pub Date : 2010-12-10 DOI: 10.1109/MMSP.2010.5661984
K. Essafi, S. B. Jebara
The use of an adaptive feedback canceller (AFC) based on signals pre-whitening/filtering in hearing aids seems very attractive, since it limits desired signal degradation when amplification gain is increased. In this paper, we present a comparative assessment of the performances of some existing pre-whitening decorrelation based methods. The used criteria consider the adaptive filter performances, the system stability and the speech quality in terms of distortion and oscillation. Results show that the method including the loudspeaker pre-whitening and the microphone filtering is the best. Moreover, the use of an adequate algorithm for adaptive pre-whitener based on the minimization of a criterion which considers the inter-correlation between the pre-whitened loudspeaker and the filtered microphone signals improves more the performances.
在助听器中使用基于信号预白化/滤波的自适应反馈抵消器(AFC)似乎非常有吸引力,因为当放大增益增加时,它限制了期望的信号退化。本文对几种基于预白化去相关方法的性能进行了比较评价。所使用的标准考虑了自适应滤波器性能、系统稳定性以及失真和振荡方面的语音质量。结果表明,结合扬声器预白化和麦克风滤波的方法效果最好。此外,采用适当的基于最小化准则的自适应预白器算法,该准则考虑了预白后的扬声器和滤波后的麦克风信号之间的相互关系,从而提高了性能。
{"title":"A comparative study between different pre-whitening decorrelation based acoustic feedback cancellers","authors":"K. Essafi, S. B. Jebara","doi":"10.1109/MMSP.2010.5661984","DOIUrl":"https://doi.org/10.1109/MMSP.2010.5661984","url":null,"abstract":"The use of an adaptive feedback canceller (AFC) based on signals pre-whitening/filtering in hearing aids seems very attractive, since it limits desired signal degradation when amplification gain is increased. In this paper, we present a comparative assessment of the performances of some existing pre-whitening decorrelation based methods. The used criteria consider the adaptive filter performances, the system stability and the speech quality in terms of distortion and oscillation. Results show that the method including the loudspeaker pre-whitening and the microphone filtering is the best. Moreover, the use of an adequate algorithm for adaptive pre-whitener based on the minimization of a criterion which considers the inter-correlation between the pre-whitened loudspeaker and the filtered microphone signals improves more the performances.","PeriodicalId":105774,"journal":{"name":"2010 IEEE International Workshop on Multimedia Signal Processing","volume":"34 5","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114111271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2010 IEEE International Workshop on Multimedia Signal Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1