{"title":"$\\varepsilon$-LAP:针对内容分发网络的轻量级自适应缓存分区方案与审慎的大小调整决策","authors":"Peng Wang;Yu Liu;Ziqi Liu;Zhelong Zhao;Ke Liu;Ke Zhou;Zhihai Huang","doi":"10.1109/TCC.2024.3420454","DOIUrl":null,"url":null,"abstract":"As dependence on Content Delivery Networks (CDNs) increases, there is a growing need for innovative solutions to optimize cache performance amid increasing traffic and complicated cache-sharing workloads. Allocating exclusive resources to applications in CDNs boosts the overall cache hit ratio (OHR), enhancing efficiency. However, the traditional method of creating the miss ratio curve (MRC) is unsuitable for CDNs due to the diverse sizes of items and the vast number of applications, leading to high computational overhead and performance inconsistency. To tackle this issue, we propose a \n<u>l</u>\nightweight and \n<u>a</u>\ndaptive cache \n<u>p</u>\nartitioning scheme called \n<inline-formula><tex-math>$\\varepsilon$</tex-math></inline-formula>\n-LAP. This scheme uses a corresponding shadow cache for each partition and sorts them based on the average hit numbers on the granularity unit in the shadow caches. During partition resizing, \n<inline-formula><tex-math>$\\varepsilon$</tex-math></inline-formula>\n-LAP transfers storage capacity, measured in units of granularity, from the \n<inline-formula><tex-math>$(N-k+1)$</tex-math></inline-formula>\n-th (\n<inline-formula><tex-math>$k\\leq \\frac{N}{2}$</tex-math></inline-formula>\n) partition to the \n<inline-formula><tex-math>$k$</tex-math></inline-formula>\n-th partition. A learning threshold parameter, i.e., \n<inline-formula><tex-math>$\\varepsilon$</tex-math></inline-formula>\n, is also introduced to prudently determine when to resize partitions, improving caching efficiency. This can eliminate about 96.8% of unnecessary partition resizing without compromising performance. \n<inline-formula><tex-math>$\\varepsilon$</tex-math></inline-formula>\n-LAP, when deployed in \n<i>PicCloud</i>\n at \n<i>Tencent</i>\n, improved OHR by 9.34% and reduced the average user access latency by 12.5 ms. Experimental results show that \n<inline-formula><tex-math>$\\varepsilon$</tex-math></inline-formula>\n-LAP outperforms other cache partitioning schemes in terms of both OHR and access latency, and it effectively adapts to workload variations.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 3","pages":"942-953"},"PeriodicalIF":5.3000,"publicationDate":"2024-06-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"$\\\\varepsilon$ɛ-LAP: A Lightweight and Adaptive Cache Partitioning Scheme With Prudent Resizing Decisions for Content Delivery Networks\",\"authors\":\"Peng Wang;Yu Liu;Ziqi Liu;Zhelong Zhao;Ke Liu;Ke Zhou;Zhihai Huang\",\"doi\":\"10.1109/TCC.2024.3420454\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"As dependence on Content Delivery Networks (CDNs) increases, there is a growing need for innovative solutions to optimize cache performance amid increasing traffic and complicated cache-sharing workloads. Allocating exclusive resources to applications in CDNs boosts the overall cache hit ratio (OHR), enhancing efficiency. However, the traditional method of creating the miss ratio curve (MRC) is unsuitable for CDNs due to the diverse sizes of items and the vast number of applications, leading to high computational overhead and performance inconsistency. To tackle this issue, we propose a \\n<u>l</u>\\nightweight and \\n<u>a</u>\\ndaptive cache \\n<u>p</u>\\nartitioning scheme called \\n<inline-formula><tex-math>$\\\\varepsilon$</tex-math></inline-formula>\\n-LAP. This scheme uses a corresponding shadow cache for each partition and sorts them based on the average hit numbers on the granularity unit in the shadow caches. During partition resizing, \\n<inline-formula><tex-math>$\\\\varepsilon$</tex-math></inline-formula>\\n-LAP transfers storage capacity, measured in units of granularity, from the \\n<inline-formula><tex-math>$(N-k+1)$</tex-math></inline-formula>\\n-th (\\n<inline-formula><tex-math>$k\\\\leq \\\\frac{N}{2}$</tex-math></inline-formula>\\n) partition to the \\n<inline-formula><tex-math>$k$</tex-math></inline-formula>\\n-th partition. A learning threshold parameter, i.e., \\n<inline-formula><tex-math>$\\\\varepsilon$</tex-math></inline-formula>\\n, is also introduced to prudently determine when to resize partitions, improving caching efficiency. This can eliminate about 96.8% of unnecessary partition resizing without compromising performance. \\n<inline-formula><tex-math>$\\\\varepsilon$</tex-math></inline-formula>\\n-LAP, when deployed in \\n<i>PicCloud</i>\\n at \\n<i>Tencent</i>\\n, improved OHR by 9.34% and reduced the average user access latency by 12.5 ms. Experimental results show that \\n<inline-formula><tex-math>$\\\\varepsilon$</tex-math></inline-formula>\\n-LAP outperforms other cache partitioning schemes in terms of both OHR and access latency, and it effectively adapts to workload variations.\",\"PeriodicalId\":13202,\"journal\":{\"name\":\"IEEE Transactions on Cloud Computing\",\"volume\":\"12 3\",\"pages\":\"942-953\"},\"PeriodicalIF\":5.3000,\"publicationDate\":\"2024-06-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Cloud Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10577125/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Cloud Computing","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10577125/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
$\varepsilon$ɛ-LAP: A Lightweight and Adaptive Cache Partitioning Scheme With Prudent Resizing Decisions for Content Delivery Networks
As dependence on Content Delivery Networks (CDNs) increases, there is a growing need for innovative solutions to optimize cache performance amid increasing traffic and complicated cache-sharing workloads. Allocating exclusive resources to applications in CDNs boosts the overall cache hit ratio (OHR), enhancing efficiency. However, the traditional method of creating the miss ratio curve (MRC) is unsuitable for CDNs due to the diverse sizes of items and the vast number of applications, leading to high computational overhead and performance inconsistency. To tackle this issue, we propose a
l
ightweight and
a
daptive cache
p
artitioning scheme called
$\varepsilon$
-LAP. This scheme uses a corresponding shadow cache for each partition and sorts them based on the average hit numbers on the granularity unit in the shadow caches. During partition resizing,
$\varepsilon$
-LAP transfers storage capacity, measured in units of granularity, from the
$(N-k+1)$
-th (
$k\leq \frac{N}{2}$
) partition to the
$k$
-th partition. A learning threshold parameter, i.e.,
$\varepsilon$
, is also introduced to prudently determine when to resize partitions, improving caching efficiency. This can eliminate about 96.8% of unnecessary partition resizing without compromising performance.
$\varepsilon$
-LAP, when deployed in
PicCloud
at
Tencent
, improved OHR by 9.34% and reduced the average user access latency by 12.5 ms. Experimental results show that
$\varepsilon$
-LAP outperforms other cache partitioning schemes in terms of both OHR and access latency, and it effectively adapts to workload variations.
期刊介绍:
The IEEE Transactions on Cloud Computing (TCC) is dedicated to the multidisciplinary field of cloud computing. It is committed to the publication of articles that present innovative research ideas, application results, and case studies in cloud computing, focusing on key technical issues related to theory, algorithms, systems, applications, and performance.