An Online Gradient-Based Caching Policy with Logarithmic Complexity and Regret Guarantees

Damiano Carra, Giovanni Neglia
{"title":"An Online Gradient-Based Caching Policy with Logarithmic Complexity and Regret Guarantees","authors":"Damiano Carra, Giovanni Neglia","doi":"arxiv-2405.01263","DOIUrl":null,"url":null,"abstract":"The commonly used caching policies, such as LRU or LFU, exhibit optimal\nperformance only for specific traffic patterns. Even advanced Machine\nLearning-based methods, which detect patterns in historical request data,\nstruggle when future requests deviate from past trends. Recently, a new class\nof policies has emerged that makes no assumptions about the request arrival\nprocess. These algorithms solve an online optimization problem, enabling\ncontinuous adaptation to the context. They offer theoretical guarantees on the\nregret metric, which is the gap between the gain of the online policy and the\ngain of the optimal static cache allocation in hindsight. Nevertheless, the\nhigh computational complexity of these solutions hinders their practical\nadoption. In this study, we introduce a groundbreaking gradient-based online\ncaching policy, the first to achieve logarithmic computational complexity\nrelative to catalog size along with regret guarantees. This means our algorithm\ncan efficiently handle large-scale data while minimizing the performance gap\nbetween real-time decisions and optimal hindsight choices. As requests arrive,\nour policy dynamically adjusts the probabilities of including items in the\ncache, which drive cache update decisions. Our algorithm's streamlined\ncomplexity is a key advantage, enabling its application to real-world traces\nfeaturing millions of requests and items. This is a significant achievement, as\ntraces of this scale have been out of reach for existing policies with regret\nguarantees. To the best of our knowledge, our experimental results show for the\nfirst time that the regret guarantees of gradient-based caching policies bring\nsignificant benefits in scenarios of practical interest.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"837 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Operating Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.01263","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The commonly used caching policies, such as LRU or LFU, exhibit optimal performance only for specific traffic patterns. Even advanced Machine Learning-based methods, which detect patterns in historical request data, struggle when future requests deviate from past trends. Recently, a new class of policies has emerged that makes no assumptions about the request arrival process. These algorithms solve an online optimization problem, enabling continuous adaptation to the context. They offer theoretical guarantees on the regret metric, which is the gap between the gain of the online policy and the gain of the optimal static cache allocation in hindsight. Nevertheless, the high computational complexity of these solutions hinders their practical adoption. In this study, we introduce a groundbreaking gradient-based online caching policy, the first to achieve logarithmic computational complexity relative to catalog size along with regret guarantees. This means our algorithm can efficiently handle large-scale data while minimizing the performance gap between real-time decisions and optimal hindsight choices. As requests arrive, our policy dynamically adjusts the probabilities of including items in the cache, which drive cache update decisions. Our algorithm's streamlined complexity is a key advantage, enabling its application to real-world traces featuring millions of requests and items. This is a significant achievement, as traces of this scale have been out of reach for existing policies with regret guarantees. To the best of our knowledge, our experimental results show for the first time that the regret guarantees of gradient-based caching policies bring significant benefits in scenarios of practical interest.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
具有对数复杂性和遗憾保证的基于梯度的在线缓存策略
常用的缓存策略,如 LRU 或 LFU,仅在特定流量模式下表现出最佳性能。即使是先进的基于机器学习的方法,也只能检测历史请求数据中的模式,当未来的请求偏离过去的趋势时,这种方法就会陷入困境。最近,出现了一类新的策略,对请求到达过程不做任何假设。这些算法解决的是一个在线优化问题,能够不断适应环境。它们从理论上保证了 "遗憾度量",即在线策略收益与事后最优静态缓存分配收益之间的差距。然而,这些解决方案的计算复杂度很高,阻碍了它们的实际应用。在本研究中,我们引入了一种开创性的基于梯度的在线缓存策略,这是第一个实现了相对于目录大小的对数计算复杂度和后悔保证的策略。这意味着我们的算法可以高效处理大规模数据,同时最大限度地缩小实时决策与事后最优选择之间的性能差距。当请求到达时,我们的策略会动态调整缓存中包含项目的概率,从而驱动缓存更新决策。我们的算法具有简化复杂性的关键优势,使其能够应用于包含数百万个请求和项目的实际跟踪。这是一项重大成就,因为对于现有的后悔保证策略来说,这种规模的跟踪是遥不可及的。据我们所知,我们的实验结果首次表明,基于梯度的遗憾保证缓存策略在实际应用场景中具有显著优势。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Analysis of Synchronization Mechanisms in Operating Systems Skip TLB flushes for reused pages within mmap's eBPF-mm: Userspace-guided memory management in Linux with eBPF BULKHEAD: Secure, Scalable, and Efficient Kernel Compartmentalization with PKS Rethinking Programmed I/O for Fast Devices, Cheap Cores, and Coherent Interconnects
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1