随机和量子计算的累积内存下界

P. Beame, Niels Kornerup
{"title":"随机和量子计算的累积内存下界","authors":"P. Beame, Niels Kornerup","doi":"10.4230/LIPIcs.ICALP.2023.17","DOIUrl":null,"url":null,"abstract":"Cumulative memory -- the sum of space used per step over the duration of a computation -- is a fine-grained measure of time-space complexity that was introduced to analyze cryptographic applications like password hashing. It is a more accurate cost measure for algorithms that have infrequent spikes in memory usage and are run in environments such as cloud computing that allow dynamic allocation and de-allocation of resources during execution, or when many multiple instances of an algorithm are interleaved in parallel. We prove the first lower bounds on cumulative memory complexity for both sequential classical computation and quantum circuits. Moreover, we develop general paradigms for bounding cumulative memory complexity inspired by the standard paradigms for proving time-space tradeoff lower bounds that can only lower bound the maximum space used during an execution. The resulting lower bounds on cumulative memory that we obtain are just as strong as the best time-space tradeoff lower bounds, which are very often known to be tight. Although previous results for pebbling and random oracle models have yielded time-space tradeoff lower bounds larger than the cumulative memory complexity, our results show that in general computational models such separations cannot follow from known lower bound techniques and are not true for many functions. Among many possible applications of our general methods, we show that any classical sorting algorithm with success probability at least $1/\\text{poly}(n)$ requires cumulative memory $\\tilde \\Omega(n^2)$, any classical matrix multiplication algorithm requires cumulative memory $\\Omega(n^6/T)$, any quantum sorting circuit requires cumulative memory $\\Omega(n^3/T)$, and any quantum circuit that finds $k$ disjoint collisions in a random function requires cumulative memory $\\Omega(k^3n/T^2)$.","PeriodicalId":11639,"journal":{"name":"Electron. Colloquium Comput. Complex.","volume":"1 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Cumulative Memory Lower Bounds for Randomized and Quantum Computation\",\"authors\":\"P. Beame, Niels Kornerup\",\"doi\":\"10.4230/LIPIcs.ICALP.2023.17\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Cumulative memory -- the sum of space used per step over the duration of a computation -- is a fine-grained measure of time-space complexity that was introduced to analyze cryptographic applications like password hashing. It is a more accurate cost measure for algorithms that have infrequent spikes in memory usage and are run in environments such as cloud computing that allow dynamic allocation and de-allocation of resources during execution, or when many multiple instances of an algorithm are interleaved in parallel. We prove the first lower bounds on cumulative memory complexity for both sequential classical computation and quantum circuits. Moreover, we develop general paradigms for bounding cumulative memory complexity inspired by the standard paradigms for proving time-space tradeoff lower bounds that can only lower bound the maximum space used during an execution. The resulting lower bounds on cumulative memory that we obtain are just as strong as the best time-space tradeoff lower bounds, which are very often known to be tight. Although previous results for pebbling and random oracle models have yielded time-space tradeoff lower bounds larger than the cumulative memory complexity, our results show that in general computational models such separations cannot follow from known lower bound techniques and are not true for many functions. Among many possible applications of our general methods, we show that any classical sorting algorithm with success probability at least $1/\\\\text{poly}(n)$ requires cumulative memory $\\\\tilde \\\\Omega(n^2)$, any classical matrix multiplication algorithm requires cumulative memory $\\\\Omega(n^6/T)$, any quantum sorting circuit requires cumulative memory $\\\\Omega(n^3/T)$, and any quantum circuit that finds $k$ disjoint collisions in a random function requires cumulative memory $\\\\Omega(k^3n/T^2)$.\",\"PeriodicalId\":11639,\"journal\":{\"name\":\"Electron. Colloquium Comput. Complex.\",\"volume\":\"1 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Electron. Colloquium Comput. Complex.\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.4230/LIPIcs.ICALP.2023.17\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Electron. Colloquium Comput. Complex.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.4230/LIPIcs.ICALP.2023.17","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

累积内存——计算过程中每一步使用的空间总和——是一种细粒度的时空复杂性度量,用于分析密码散列等加密应用程序。对于那些内存使用量很少出现峰值的算法,以及在云计算等允许在执行期间动态分配和取消资源分配的环境中运行的算法,或者当一个算法的多个实例并行交错时,它是一种更准确的成本度量。我们证明了顺序经典计算和量子电路的累积存储器复杂度的第一个下界。此外,我们开发了限制累积内存复杂性的通用范例,其灵感来自于证明时间-空间权衡下界的标准范例,该下界只能在执行期间使用的最大空间下界。我们得到的累积内存的下界和最佳的时间-空间折衷下界一样强,后者通常都很紧。尽管先前的鹅卵石和随机oracle模型的结果已经产生了大于累积内存复杂性的时空权衡下界,但我们的结果表明,在一般的计算模型中,这种分离不能遵循已知的下界技术,并且对于许多函数来说并不正确。在我们的一般方法的许多可能应用中,我们表明,任何成功概率至少$1/\text{poly}(n)$的经典排序算法都需要累积内存$\tilde \Omega(n^2)$,任何经典矩阵乘法算法都需要累积内存$\Omega(n^6/T)$,任何量子排序电路都需要累积内存$\Omega(n^3/T)$,任何在随机函数中发现$k$不相交碰撞的量子电路都需要累积内存$\Omega(k^3n/T^2)$。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Cumulative Memory Lower Bounds for Randomized and Quantum Computation
Cumulative memory -- the sum of space used per step over the duration of a computation -- is a fine-grained measure of time-space complexity that was introduced to analyze cryptographic applications like password hashing. It is a more accurate cost measure for algorithms that have infrequent spikes in memory usage and are run in environments such as cloud computing that allow dynamic allocation and de-allocation of resources during execution, or when many multiple instances of an algorithm are interleaved in parallel. We prove the first lower bounds on cumulative memory complexity for both sequential classical computation and quantum circuits. Moreover, we develop general paradigms for bounding cumulative memory complexity inspired by the standard paradigms for proving time-space tradeoff lower bounds that can only lower bound the maximum space used during an execution. The resulting lower bounds on cumulative memory that we obtain are just as strong as the best time-space tradeoff lower bounds, which are very often known to be tight. Although previous results for pebbling and random oracle models have yielded time-space tradeoff lower bounds larger than the cumulative memory complexity, our results show that in general computational models such separations cannot follow from known lower bound techniques and are not true for many functions. Among many possible applications of our general methods, we show that any classical sorting algorithm with success probability at least $1/\text{poly}(n)$ requires cumulative memory $\tilde \Omega(n^2)$, any classical matrix multiplication algorithm requires cumulative memory $\Omega(n^6/T)$, any quantum sorting circuit requires cumulative memory $\Omega(n^3/T)$, and any quantum circuit that finds $k$ disjoint collisions in a random function requires cumulative memory $\Omega(k^3n/T^2)$.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Dependency schemes in CDCL-based QBF solving: a proof-theoretic study On blocky ranks of matrices Fractional Linear Matroid Matching is in quasi-NC Aaronson-Ambainis Conjecture Is True For Random Restrictions Optimal Pseudorandom Generators for Low-Degree Polynomials Over Moderately Large Fields
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1