{"title":"对slp压缩文档的扳手评估","authors":"Markus L. Schmid, Nicole Schweikardt","doi":"10.1145/3452021.3458325","DOIUrl":null,"url":null,"abstract":"We consider the problem of evaluating regular spanners over compressed documents, i.e., we wish to solve evaluation tasks directly on the compressed data, without decompression. As compressed forms of the documents we use straight-line programs (SLPs) --- a lossless compression scheme for textual data widely used in different areas of theoretical computer science and particularly well-suited for algorithmics on compressed data. In data complexity, our results are as follows. For a regular spanner M and an SLP $\\mathcalS $ of size $\\mathbfs $ that represents a document D, we can solve the tasks of model checking and of checking non-emptiness in time $O(\\mathbfs )$. Computing the set $łlbracket M \\rrbracket(D)$ of all span-tuples extracted from D can be done in time $Ø(\\mathbfs |łlbracket M \\rrbracket(D)|)$, and enumeration of $łlbracket M \\rrbracket(D)$ can be done with linear preprocessing $O(\\mathbfs )$ and a delay of $O(depth\\mathcalS )$, where $depth\\mathcalS $ is the depth of $\\mathcalS $'s derivation tree. Note that $\\mathbfs $ can be exponentially smaller than the document's size $|D|$; and, due to known balancing results for SLPs, we can always assume that $depth\\mathcalS = O(log(|D|))$ independent of D's compressibility. Hence, our enumeration algorithm has a delay logarithmic in the size of the non-compressed data and a preprocessing time that is at best (i.e., in the case of highly compressible documents) also logarithmic, but at worst still linear. Therefore, in a big-data perspective, our enumeration algorithm for SLP-compressed documents may nevertheless beat the known linear preprocessing and constant delay algorithms for non-compressed documents.","PeriodicalId":405398,"journal":{"name":"Proceedings of the 40th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"Spanner Evaluation over SLP-Compressed Documents\",\"authors\":\"Markus L. Schmid, Nicole Schweikardt\",\"doi\":\"10.1145/3452021.3458325\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We consider the problem of evaluating regular spanners over compressed documents, i.e., we wish to solve evaluation tasks directly on the compressed data, without decompression. As compressed forms of the documents we use straight-line programs (SLPs) --- a lossless compression scheme for textual data widely used in different areas of theoretical computer science and particularly well-suited for algorithmics on compressed data. In data complexity, our results are as follows. For a regular spanner M and an SLP $\\\\mathcalS $ of size $\\\\mathbfs $ that represents a document D, we can solve the tasks of model checking and of checking non-emptiness in time $O(\\\\mathbfs )$. Computing the set $łlbracket M \\\\rrbracket(D)$ of all span-tuples extracted from D can be done in time $Ø(\\\\mathbfs |łlbracket M \\\\rrbracket(D)|)$, and enumeration of $łlbracket M \\\\rrbracket(D)$ can be done with linear preprocessing $O(\\\\mathbfs )$ and a delay of $O(depth\\\\mathcalS )$, where $depth\\\\mathcalS $ is the depth of $\\\\mathcalS $'s derivation tree. Note that $\\\\mathbfs $ can be exponentially smaller than the document's size $|D|$; and, due to known balancing results for SLPs, we can always assume that $depth\\\\mathcalS = O(log(|D|))$ independent of D's compressibility. Hence, our enumeration algorithm has a delay logarithmic in the size of the non-compressed data and a preprocessing time that is at best (i.e., in the case of highly compressible documents) also logarithmic, but at worst still linear. Therefore, in a big-data perspective, our enumeration algorithm for SLP-compressed documents may nevertheless beat the known linear preprocessing and constant delay algorithms for non-compressed documents.\",\"PeriodicalId\":405398,\"journal\":{\"name\":\"Proceedings of the 40th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems\",\"volume\":\"9 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-01-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 40th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3452021.3458325\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 40th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3452021.3458325","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8
摘要
我们考虑在压缩文档上评估常规生成器的问题,即,我们希望直接在压缩数据上解决评估任务,而不需要解压缩。作为文件的压缩形式,我们使用直线程序(slp)——一种文本数据的无损压缩方案,广泛应用于理论计算机科学的不同领域,特别适合压缩数据的算法。在数据复杂度方面,我们的结果如下。对于一个普通的扳手M和一个大小为$\mathbfs $的SLP $\mathcal $表示文档D,我们可以解决模型检查和在时间$O(\mathbfs)$检查非空的任务。计算从D中提取的所有跨度元组的集合$łlbracket M \ rr括号(D)$可以在时间$Ø(\mathbfs |łlbracket M \ rr括号(D)|)$中完成,而$łlbracket M \ rr括号(D)$的枚举可以通过线性预处理$O(\mathbfs)$和延迟$O(depth\mathcalS)$来完成,其中$depth\mathcalS $是$\mathcalS $的派生树的深度。注意$\mathbfs $可以比文档的大小指数小$|D|$;并且,由于已知slp的平衡结果,我们总是可以假设$depth\mathcal = O(log(|D|))$独立于D的可压缩性。因此,我们的枚举算法在未压缩数据的大小上具有对数级的延迟,而预处理时间在最好的情况下(即在高度可压缩的文档的情况下)也是对数级的,但在最坏的情况下仍然是线性的。因此,从大数据的角度来看,我们针对slp压缩文档的枚举算法可能优于针对非压缩文档的已知线性预处理和恒定延迟算法。
We consider the problem of evaluating regular spanners over compressed documents, i.e., we wish to solve evaluation tasks directly on the compressed data, without decompression. As compressed forms of the documents we use straight-line programs (SLPs) --- a lossless compression scheme for textual data widely used in different areas of theoretical computer science and particularly well-suited for algorithmics on compressed data. In data complexity, our results are as follows. For a regular spanner M and an SLP $\mathcalS $ of size $\mathbfs $ that represents a document D, we can solve the tasks of model checking and of checking non-emptiness in time $O(\mathbfs )$. Computing the set $łlbracket M \rrbracket(D)$ of all span-tuples extracted from D can be done in time $Ø(\mathbfs |łlbracket M \rrbracket(D)|)$, and enumeration of $łlbracket M \rrbracket(D)$ can be done with linear preprocessing $O(\mathbfs )$ and a delay of $O(depth\mathcalS )$, where $depth\mathcalS $ is the depth of $\mathcalS $'s derivation tree. Note that $\mathbfs $ can be exponentially smaller than the document's size $|D|$; and, due to known balancing results for SLPs, we can always assume that $depth\mathcalS = O(log(|D|))$ independent of D's compressibility. Hence, our enumeration algorithm has a delay logarithmic in the size of the non-compressed data and a preprocessing time that is at best (i.e., in the case of highly compressible documents) also logarithmic, but at worst still linear. Therefore, in a big-data perspective, our enumeration algorithm for SLP-compressed documents may nevertheless beat the known linear preprocessing and constant delay algorithms for non-compressed documents.