High-Quality and Low-Memory-Footprint Progressive Decoding of Large-Scale Particle Data

D. Hoang, H. Bhatia, P. Lindstrom, Valerio Pascucci
{"title":"High-Quality and Low-Memory-Footprint Progressive Decoding of Large-Scale Particle Data","authors":"D. Hoang, H. Bhatia, P. Lindstrom, Valerio Pascucci","doi":"10.1109/LDAV53230.2021.00011","DOIUrl":null,"url":null,"abstract":"Particle representations are used often in large-scale simulations and observations, frequently creating datasets containing several millions of particles or more. Due to their sheer size, such datasets are difficult to store, transfer, and analyze efficiently. Data compression is a promising solution; however, effective approaches to compress particle data are lacking and no community-standard and accepted techniques exist. Current techniques are designed either to compress small data very well but require high computational resources when applied to large data, or to work with large data but without a focus on compression, resulting in low reconstruction quality per bit stored. In this paper, we present innovations targeting tree-based particle compression approaches that improve the tradeoff between high quality and low memory-footprint for compression and decompression of large particle datasets. Inspired by the lazy wavelet transform, we introduce a new way of partitioning space, which allows a low-cost depth-first traversal of a particle hierarchy to cover the space broadly. We also devise novel data-adaptive traversal orders that significantly reduce reconstruction error compared to traditional data-agnostic orders such as breadth-first and depth-first traversals. The new partitioning and traversal schemes are used to build novel particle hierarchies that can be traversed with asymptotically constant memory footprint while incurring low reconstruction error. Our solution to encoding and (lossy) decoding of large particle data is a flexible block-based hierarchy that supports progressive, random-access, and error-driven decoding, where error heuristics can be supplied by the user. Finally, through extensive experimentation, we demonstrate the efficacy and the flexibility of the proposed techniques when combined as well as when used independently with existing approaches on a wide range of scientific particle datasets.","PeriodicalId":441438,"journal":{"name":"2021 IEEE 11th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"192 3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 11th Symposium on Large Data Analysis and Visualization (LDAV)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/LDAV53230.2021.00011","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

Particle representations are used often in large-scale simulations and observations, frequently creating datasets containing several millions of particles or more. Due to their sheer size, such datasets are difficult to store, transfer, and analyze efficiently. Data compression is a promising solution; however, effective approaches to compress particle data are lacking and no community-standard and accepted techniques exist. Current techniques are designed either to compress small data very well but require high computational resources when applied to large data, or to work with large data but without a focus on compression, resulting in low reconstruction quality per bit stored. In this paper, we present innovations targeting tree-based particle compression approaches that improve the tradeoff between high quality and low memory-footprint for compression and decompression of large particle datasets. Inspired by the lazy wavelet transform, we introduce a new way of partitioning space, which allows a low-cost depth-first traversal of a particle hierarchy to cover the space broadly. We also devise novel data-adaptive traversal orders that significantly reduce reconstruction error compared to traditional data-agnostic orders such as breadth-first and depth-first traversals. The new partitioning and traversal schemes are used to build novel particle hierarchies that can be traversed with asymptotically constant memory footprint while incurring low reconstruction error. Our solution to encoding and (lossy) decoding of large particle data is a flexible block-based hierarchy that supports progressive, random-access, and error-driven decoding, where error heuristics can be supplied by the user. Finally, through extensive experimentation, we demonstrate the efficacy and the flexibility of the proposed techniques when combined as well as when used independently with existing approaches on a wide range of scientific particle datasets.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
大规模粒子数据的高质量和低内存占用渐进解码
粒子表示通常用于大规模模拟和观测,经常创建包含数百万或更多粒子的数据集。由于其庞大的规模,这些数据集很难有效地存储、传输和分析。数据压缩是一个很有前途的解决方案;然而,目前缺乏有效的压缩粒子数据的方法,也没有统一的标准和公认的技术。目前的技术要么是为了压缩小数据而设计的,但在应用于大数据时需要大量的计算资源,要么是为了处理大数据而不关注压缩,导致每比特存储的重构质量很低。在本文中,我们提出了针对基于树的粒子压缩方法的创新,这些方法在压缩和解压大型粒子数据集时改善了高质量和低内存占用之间的权衡。受惰性小波变换的启发,我们引入了一种新的空间划分方法,该方法允许对粒子层次结构进行低成本的深度优先遍历,以广泛地覆盖空间。我们还设计了新的数据自适应遍历顺序,与传统的数据不可知顺序(如宽度优先和深度优先遍历)相比,它显著减少了重构误差。新的分区和遍历方案用于构建新的粒子层次结构,该结构可以在具有渐进恒定内存占用的情况下遍历,同时产生低重构错误。我们对大粒子数据的编码和(有损)解码的解决方案是一个灵活的基于块的层次结构,它支持渐进式、随机访问和错误驱动的解码,其中错误启发式可以由用户提供。最后,通过广泛的实验,我们证明了所提出的技术在结合以及与广泛的科学粒子数据集上的现有方法独立使用时的有效性和灵活性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
IExchange: Asynchronous Communication and Termination Detection for Iterative Algorithms Parameter Analysis and Contrail Detection of Aircraft Engine Simulations An Entropy-Based Approach for Identifying User-Preferred Camera Positions Portable and Composable Flow Graphs for In Situ Analytics Lossy Compression for Visualization of Atmospheric Data
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1