首页 > 最新文献

2018 Picture Coding Symposium (PCS)最新文献

英文 中文
Signal and Loss Geometry Aware Frequency Selective Extrapolation for Error Concealment 误差隐藏的信号和损耗几何感知频率选择外推
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456259
Nils Genser, Jürgen Seiler, Franz Schilling, André Kaup
The concealment of errors is an important task in image and video signal processing. Often, complex models are calculated to reconstruct the missing samples, which results in a long computation time. One method that achieves a very high reconstruction quality, but demands a moderate computational complexity only, is the block based Frequency Selective Extrapolation. Nevertheless, the reconstruction of a Full HD image can still take several minutes depending on the error pattern. To accelerate the computation, a novel algorithm is introduced in this paper that analyzes the adjacent, undistorted samples and optimizes the reconstruction parameters accordingly. Moreover, the analyzation is further used to adapt the partitioning of the blocks and the processing order. Similar to modern video codecs, e.g., High Efficiency Video Coding, a content based partitioning and processing is proposed as it takes the signal characteristics into account. Thus, the novel algorithm is on average four times faster than the state-of-the-art method and up to $25times $ quicker at best, while achieving a slightly higher reconstruction quality as well.
错误隐藏是图像和视频信号处理中的一项重要任务。通常需要计算复杂的模型来重建缺失的样本,这导致计算时间很长。基于分块的频率选择外推法是一种实现高质量重构的方法,但只需要适度的计算复杂度。然而,全高清图像的重建仍然需要几分钟,这取决于错误模式。为了加快计算速度,本文介绍了一种新的算法,该算法对相邻的、未失真的样本进行分析,并相应地优化重建参数。此外,还进一步利用分析来调整块的划分和处理顺序。与高效视频编码(High Efficiency video Coding)等现代视频编解码器类似,本文提出了一种考虑信号特性的基于内容的分割和处理方法。因此,新算法的平均速度比最先进的方法快4倍,最多快25倍,同时实现了略高的重建质量。
{"title":"Signal and Loss Geometry Aware Frequency Selective Extrapolation for Error Concealment","authors":"Nils Genser, Jürgen Seiler, Franz Schilling, André Kaup","doi":"10.1109/PCS.2018.8456259","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456259","url":null,"abstract":"The concealment of errors is an important task in image and video signal processing. Often, complex models are calculated to reconstruct the missing samples, which results in a long computation time. One method that achieves a very high reconstruction quality, but demands a moderate computational complexity only, is the block based Frequency Selective Extrapolation. Nevertheless, the reconstruction of a Full HD image can still take several minutes depending on the error pattern. To accelerate the computation, a novel algorithm is introduced in this paper that analyzes the adjacent, undistorted samples and optimizes the reconstruction parameters accordingly. Moreover, the analyzation is further used to adapt the partitioning of the blocks and the processing order. Similar to modern video codecs, e.g., High Efficiency Video Coding, a content based partitioning and processing is proposed as it takes the signal characteristics into account. Thus, the novel algorithm is on average four times faster than the state-of-the-art method and up to $25times $ quicker at best, while achieving a slightly higher reconstruction quality as well.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127569815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
CNN-based Prediction for Lossless Coding of Photographic Images 基于cnn的图像无损编码预测
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456311
I. Schiopu, Yu Liu, A. Munteanu
The paper proposes a novel prediction paradigm in image coding based on Convolutional Neural Networks (CNN). A deep neural network is designed to provide accurate pixel-wise prediction based on a causal neighbourhood. The proposed CNN prediction method is trained on the high-activity areas in the image and it is incorporated in a lossless compression system for high-resolution photographic images. The system uses the proposed CNN-based prediction paradigm as well as LOCO-I, whereby the predictor selection is performed using a local entropy-based descriptor. The prediction errors are encoded using a CALIC-based reference codec. The experimental results show a good performance for the proposed prediction scheme compared to state-of-the-art predictors. To our knowledge, the paper is the first to introduce CNN-based prediction in image coding, and demonstrates the potential offered by machine learning methods in coding applications.
提出了一种基于卷积神经网络(CNN)的图像编码预测范式。深度神经网络被设计用于基于因果邻域提供精确的逐像素预测。提出的CNN预测方法对图像中的高活动区域进行训练,并将其纳入高分辨率摄影图像的无损压缩系统中。该系统使用提出的基于cnn的预测范式以及LOCO-I,其中预测器选择使用基于局部熵的描述符执行。使用基于calic的参考编解码器对预测误差进行编码。实验结果表明,与现有的预测器相比,所提出的预测方案具有良好的性能。据我们所知,这篇论文首次将基于cnn的预测引入到图像编码中,并展示了机器学习方法在编码应用中的潜力。
{"title":"CNN-based Prediction for Lossless Coding of Photographic Images","authors":"I. Schiopu, Yu Liu, A. Munteanu","doi":"10.1109/PCS.2018.8456311","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456311","url":null,"abstract":"The paper proposes a novel prediction paradigm in image coding based on Convolutional Neural Networks (CNN). A deep neural network is designed to provide accurate pixel-wise prediction based on a causal neighbourhood. The proposed CNN prediction method is trained on the high-activity areas in the image and it is incorporated in a lossless compression system for high-resolution photographic images. The system uses the proposed CNN-based prediction paradigm as well as LOCO-I, whereby the predictor selection is performed using a local entropy-based descriptor. The prediction errors are encoded using a CALIC-based reference codec. The experimental results show a good performance for the proposed prediction scheme compared to state-of-the-art predictors. To our knowledge, the paper is the first to introduce CNN-based prediction in image coding, and demonstrates the potential offered by machine learning methods in coding applications.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"80 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128440750","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Deep Convolutional AutoEncoder-based Lossy Image Compression 基于深度卷积自动编码器的有损图像压缩
Pub Date : 2018-04-25 DOI: 10.1109/PCS.2018.8456308
Zhengxue Cheng, Heming Sun, Masaru Takeuchi, J. Katto
Image compression has been investigated as a fundamental research topic for many decades. Recently, deep learning has achieved great success in many computer vision tasks, and is gradually being used in image compression. In this paper, we present a lossy image compression architecture, which utilizes the advantages of convolutional autoencoder (CAE) to achieve a high coding efficiency. First, we design a novel CAE architecture to replace the conventional transforms and train this CAE using a rate-distortion loss function. Second, to generate a more energy-compact representation, we utilize the principal components analysis (PCA) to rotate the feature maps produced by the CAE, and then apply the quantization and entropy coder to generate the codes. Experimental results demonstrate that our method outperforms traditional image coding algorithms, by achieving a 13.7% BD-rate decrement on the Kodak database images compared to JPEG2000. Besides, our method maintains a moderate complexity similar to JPEG2000.
图像压缩作为一个基础研究课题已经进行了几十年的研究。近年来,深度学习在许多计算机视觉任务中取得了巨大的成功,并逐渐应用于图像压缩。在本文中,我们提出了一种有损图像压缩架构,利用卷积自编码器(CAE)的优点来实现高编码效率。首先,我们设计了一种新的CAE架构来取代传统的变换,并使用率失真损失函数来训练该CAE。其次,为了生成更紧凑的能量表示,我们利用主成分分析(PCA)来旋转CAE生成的特征映射,然后应用量化和熵编码器来生成代码。实验结果表明,我们的方法优于传统的图像编码算法,与JPEG2000相比,柯达数据库图像的bd率降低了13.7%。此外,我们的方法保持了类似于JPEG2000的中等复杂度。
{"title":"Deep Convolutional AutoEncoder-based Lossy Image Compression","authors":"Zhengxue Cheng, Heming Sun, Masaru Takeuchi, J. Katto","doi":"10.1109/PCS.2018.8456308","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456308","url":null,"abstract":"Image compression has been investigated as a fundamental research topic for many decades. Recently, deep learning has achieved great success in many computer vision tasks, and is gradually being used in image compression. In this paper, we present a lossy image compression architecture, which utilizes the advantages of convolutional autoencoder (CAE) to achieve a high coding efficiency. First, we design a novel CAE architecture to replace the conventional transforms and train this CAE using a rate-distortion loss function. Second, to generate a more energy-compact representation, we utilize the principal components analysis (PCA) to rotate the feature maps produced by the CAE, and then apply the quantization and entropy coder to generate the codes. Experimental results demonstrate that our method outperforms traditional image coding algorithms, by achieving a 13.7% BD-rate decrement on the Kodak database images compared to JPEG2000. Besides, our method maintains a moderate complexity similar to JPEG2000.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123588538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 133
Efficient Nonlinear Transforms for Lossy Image Compression 有损图像压缩的有效非线性变换
Pub Date : 2018-01-31 DOI: 10.1109/PCS.2018.8456272
J. Ballé
We assess the performance of two techniques in the context of nonlinear transform coding with artificial neural networks, Sadam and GDN. Both techniques have been success- fully used in state-of-the-art image compression methods, but their performance has not been individually assessed to this point. Together, the techniques stabilize the training procedure of nonlinear image transforms and increase their capacity to approximate the (unknown) rate-distortion optimal transform functions. Besides comparing their performance to established alternatives, we detail the implementation of both methods and provide open-source code along with the paper.
我们评估了两种技术在人工神经网络(Sadam和GDN)的非线性变换编码中的性能。这两种技术都成功地应用于最先进的图像压缩方法,但它们的性能还没有单独评估到这一点。总之,这些技术稳定了非线性图像变换的训练过程,提高了它们近似(未知)率失真最优变换函数的能力。除了将它们的性能与现有的替代方法进行比较外,我们还详细介绍了这两种方法的实现,并提供了开源代码。
{"title":"Efficient Nonlinear Transforms for Lossy Image Compression","authors":"J. Ballé","doi":"10.1109/PCS.2018.8456272","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456272","url":null,"abstract":"We assess the performance of two techniques in the context of nonlinear transform coding with artificial neural networks, Sadam and GDN. Both techniques have been success- fully used in state-of-the-art image compression methods, but their performance has not been individually assessed to this point. Together, the techniques stabilize the training procedure of nonlinear image transforms and increase their capacity to approximate the (unknown) rate-distortion optimal transform functions. Besides comparing their performance to established alternatives, we detail the implementation of both methods and provide open-source code along with the paper.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126228391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 66
A Saak Transform Approach to Efficient, Scalable and Robust Handwritten Digits Recognition 一种高效、可扩展、鲁棒手写数字识别的Saak变换方法
Pub Date : 2017-10-29 DOI: 10.1109/PCS.2018.8456277
Yueru Chen, Zhuwei Xu, Shanshan Cai, Yujian Lang, C.-C. Jay Kuo
An efficient, scalable and robust approach to the handwritten digits recognition problem based on the Saak transform is proposed in this work. First, multi-stage Saak transforms are used to extract a family of joint spatial-spectral representations of input images. Then, the Saak coefficients are used as features and fed into the SVM classifier for the classification task. In order to control the size of Saak coefficients, we adopt a lossy Saak transform that uses the principal component analysis (PCA) to select a smaller set of transform kernels. The handwritten digits recognition problem is well solved by the convolutional neural network (CNN) such as the LeNet-5. We conduct a comparative study on the performance of the LeNet-5 and the Saak-transform-based solutions in terms of scalability and robustness as well as the efficiency of lossless and lossy Saak transforms under a comparable accuracy level.
本文提出了一种基于Saak变换的高效、可扩展、鲁棒的手写数字识别方法。首先,使用多阶段Saak变换提取输入图像的一系列联合空间-光谱表示。然后,将Saak系数作为特征输入到SVM分类器中进行分类任务。为了控制Saak系数的大小,我们采用有损Saak变换,该变换使用主成分分析(PCA)来选择较小的变换核集。LeNet-5等卷积神经网络(CNN)很好地解决了手写数字识别问题。我们对LeNet-5和基于Saak变换的解决方案在可扩展性和鲁棒性方面的性能以及在相当精度水平下无损和有损Saak变换的效率进行了比较研究。
{"title":"A Saak Transform Approach to Efficient, Scalable and Robust Handwritten Digits Recognition","authors":"Yueru Chen, Zhuwei Xu, Shanshan Cai, Yujian Lang, C.-C. Jay Kuo","doi":"10.1109/PCS.2018.8456277","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456277","url":null,"abstract":"An efficient, scalable and robust approach to the handwritten digits recognition problem based on the Saak transform is proposed in this work. First, multi-stage Saak transforms are used to extract a family of joint spatial-spectral representations of input images. Then, the Saak coefficients are used as features and fed into the SVM classifier for the classification task. In order to control the size of Saak coefficients, we adopt a lossy Saak transform that uses the principal component analysis (PCA) to select a smaller set of transform kernels. The handwritten digits recognition problem is well solved by the convolutional neural network (CNN) such as the LeNet-5. We conduct a comparative study on the performance of the LeNet-5 and the Saak-transform-based solutions in terms of scalability and robustness as well as the efficiency of lossless and lossy Saak transforms under a comparable accuracy level.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"220 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122521368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Generative Compression 生成压缩
Pub Date : 2017-03-04 DOI: 10.1109/PCS.2018.8456298
Shibani Santurkar, D. Budden, N. Shavit
Traditional image and video compression algorithms rely on hand-crafted encoder/decoder pairs (codecs) that lack adaptability and are agnostic to the data being compressed. We describe the concept of generative compression, the compression of data using generative models, and suggest that it is a direction worth pursuing to produce more accurate and visually pleasing reconstructions at deeper compression levels for both image and video data. We also show that generative compression is orders- of-magnitude more robust to bit errors (e.g., from noisy channels) than traditional variable-length coding schemes.
传统的图像和视频压缩算法依赖于手工制作的编码器/解码器对(编解码器),缺乏适应性,并且对被压缩的数据不可知。我们描述了生成压缩的概念,使用生成模型对数据进行压缩,并建议在更深的压缩水平上为图像和视频数据产生更准确和视觉上令人愉悦的重建是一个值得追求的方向。我们还表明,生成压缩比传统的变长编码方案对比特错误(例如,来自噪声信道)更具数量级的鲁棒性。
{"title":"Generative Compression","authors":"Shibani Santurkar, D. Budden, N. Shavit","doi":"10.1109/PCS.2018.8456298","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456298","url":null,"abstract":"Traditional image and video compression algorithms rely on hand-crafted encoder/decoder pairs (codecs) that lack adaptability and are agnostic to the data being compressed. We describe the concept of generative compression, the compression of data using generative models, and suggest that it is a direction worth pursuing to produce more accurate and visually pleasing reconstructions at deeper compression levels for both image and video data. We also show that generative compression is orders- of-magnitude more robust to bit errors (e.g., from noisy channels) than traditional variable-length coding schemes.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117351791","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 174
期刊
2018 Picture Coding Symposium (PCS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1