首页 > 最新文献

2021 IEEE International Conference on Networking, Architecture and Storage (NAS)最新文献

英文 中文
NVSwap: Latency-Aware Paging using Non-Volatile Main Memory NVSwap:使用非易失性主存的延迟感知分页
Pub Date : 2021-10-01 DOI: 10.1109/nas51552.2021.9605418
Yekang Wu, Xuechen Zhang
Page relocation (paging) from DRAM to swap devices is an important task of a virtual memory system in operating systems. Existing Linux paging mechanisms have two main deficiencies: (1) they may incur a high I/O latency due to write interference on solid-state disks and aggressive memory page reclaiming rate under high memory pressure and (2) they do not provide predictable latency bound for latency-sensitive applications because they cannot control the allocation of system resources among concurrent processes sharing swap devices.In this paper, we present the design and implementation of a latency-aware paging mechanism called NVSwap. It supports a hybrid swap space using both regular secondary storage devices (e.g., solid-state disks) and non-volatile main memory (NVMM). The design is more cost-effective than using only NVMM as swap spaces. Furthermore, NVSwap uses NVMM as a persistent paging buffer to serve the page-out requests and hide the latency of paging between the regular swap device and DRAM. It supports in-situ paging for pages in the persistent paging buffer avoiding the slow I/O path. Finally, NVSwap allows users to specify latency bounds for individual processes or a group of related processes and enforces the bounds by dynamically controlling the resource allocation of NVMM and page reclaiming rate in memory among scheduling units. We have implemented a prototype of NVSwap in the Linux kernel-4.4.241 based on Intel Optane DIMMs. Our results demonstrate that NVSwap reduces paging latency by up to 99% and provides performance guarantee and isolation among concurrent applications sharing swap devices.
从DRAM到交换设备的页面重定位(分页)是操作系统中虚拟内存系统的一项重要任务。现有的Linux分页机制有两个主要缺陷:(1)由于固态磁盘上的写干扰和高内存压力下的内存页面回收率,它们可能导致较高的I/O延迟;(2)它们不能为对延迟敏感的应用程序提供可预测的延迟绑定,因为它们不能控制共享交换设备的并发进程之间的系统资源分配。在本文中,我们介绍了一种称为NVSwap的延迟感知分页机制的设计和实现。它支持使用常规辅助存储设备(例如,固态磁盘)和非易失性主存储器(NVMM)的混合交换空间。这种设计比仅使用NVMM作为交换空间更具成本效益。此外,NVSwap使用NVMM作为持久的分页缓冲区来处理出页请求,并隐藏常规交换设备和DRAM之间的分页延迟。它支持对持久分页缓冲区中的页面进行原位分页,避免了缓慢的I/O路径。最后,NVSwap允许用户为单个进程或一组相关进程指定延迟界限,并通过动态控制NVMM的资源分配和调度单元之间内存中的页面回收率来强制执行该界限。我们在基于Intel Optane dimm的Linux内核4.4.241中实现了NVSwap的原型。我们的结果表明,NVSwap将分页延迟减少了99%,并在共享交换设备的并发应用程序之间提供了性能保证和隔离。
{"title":"NVSwap: Latency-Aware Paging using Non-Volatile Main Memory","authors":"Yekang Wu, Xuechen Zhang","doi":"10.1109/nas51552.2021.9605418","DOIUrl":"https://doi.org/10.1109/nas51552.2021.9605418","url":null,"abstract":"Page relocation (paging) from DRAM to swap devices is an important task of a virtual memory system in operating systems. Existing Linux paging mechanisms have two main deficiencies: (1) they may incur a high I/O latency due to write interference on solid-state disks and aggressive memory page reclaiming rate under high memory pressure and (2) they do not provide predictable latency bound for latency-sensitive applications because they cannot control the allocation of system resources among concurrent processes sharing swap devices.In this paper, we present the design and implementation of a latency-aware paging mechanism called NVSwap. It supports a hybrid swap space using both regular secondary storage devices (e.g., solid-state disks) and non-volatile main memory (NVMM). The design is more cost-effective than using only NVMM as swap spaces. Furthermore, NVSwap uses NVMM as a persistent paging buffer to serve the page-out requests and hide the latency of paging between the regular swap device and DRAM. It supports in-situ paging for pages in the persistent paging buffer avoiding the slow I/O path. Finally, NVSwap allows users to specify latency bounds for individual processes or a group of related processes and enforces the bounds by dynamically controlling the resource allocation of NVMM and page reclaiming rate in memory among scheduling units. We have implemented a prototype of NVSwap in the Linux kernel-4.4.241 based on Intel Optane DIMMs. Our results demonstrate that NVSwap reduces paging latency by up to 99% and provides performance guarantee and isolation among concurrent applications sharing swap devices.","PeriodicalId":135930,"journal":{"name":"2021 IEEE International Conference on Networking, Architecture and Storage (NAS)","volume":"160 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114173714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
[Copyright notice] (版权)
Pub Date : 2021-10-01 DOI: 10.1109/nas51552.2021.9605439
{"title":"[Copyright notice]","authors":"","doi":"10.1109/nas51552.2021.9605439","DOIUrl":"https://doi.org/10.1109/nas51552.2021.9605439","url":null,"abstract":"","PeriodicalId":135930,"journal":{"name":"2021 IEEE International Conference on Networking, Architecture and Storage (NAS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128911315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Characterizing AI Model Inference Applications Running in the SGX Environment 描述在SGX环境中运行的AI模型推理应用程序
Pub Date : 2021-10-01 DOI: 10.1109/nas51552.2021.9605445
Shixiong Jing, Qinkun Bao, Pei Wang, Xulong Tang, Dinghao Wu
Intel Software Guard Extensions (SGX) is a set of extensions built into Intel CPUs for the trusted computation. It creates a hardware-assisted secure container, within which programs are protected from data leakage and data manipulations by privileged software and hypervisors. With the trend that more and more machine learning based programs are moving to cloud computing, SGX can be used in cloud-based Machine Learning applications to protect user data from malicious privileged programs.However, applications running in SGX suffer from several overheads, including frequent context switching, memory page encryption/decryption, and memory page swapping, which significantly degrade the execution efficiency. In this paper, we aim to i) comprehensively explore the execution of general AI applications running on SGX, ii) systematically characterize the data reuses at both page granularity and cacheline granularity, and iii) provide optimization insights for efficient deployment of machine learning based applications on SGX. To the best of our knowledge, our work is the first to study machine learning applications on SGX and explore the potential of data reuses to reduce the runtime overheads in SGX.
Intel Software Guard Extensions (SGX)是一组内置于Intel cpu中的扩展,用于可信计算。它创建了一个硬件辅助的安全容器,在其中保护程序免受数据泄漏和特权软件和管理程序的数据操作。随着越来越多的基于机器学习的程序转向云计算的趋势,SGX可以用于基于云的机器学习应用程序,以保护用户数据免受恶意特权程序的侵害。但是,在SGX中运行的应用程序有一些开销,包括频繁的上下文切换、内存页加密/解密和内存页交换,这些开销会显著降低执行效率。在本文中,我们的目标是i)全面探索在SGX上运行的通用AI应用程序的执行,ii)系统地描述页面粒度和缓存粒度的数据重用,以及iii)为在SGX上有效部署基于机器学习的应用程序提供优化见解。据我们所知,我们的工作是第一个研究SGX上的机器学习应用程序,并探索数据重用的潜力,以减少SGX的运行时开销。
{"title":"Characterizing AI Model Inference Applications Running in the SGX Environment","authors":"Shixiong Jing, Qinkun Bao, Pei Wang, Xulong Tang, Dinghao Wu","doi":"10.1109/nas51552.2021.9605445","DOIUrl":"https://doi.org/10.1109/nas51552.2021.9605445","url":null,"abstract":"Intel Software Guard Extensions (SGX) is a set of extensions built into Intel CPUs for the trusted computation. It creates a hardware-assisted secure container, within which programs are protected from data leakage and data manipulations by privileged software and hypervisors. With the trend that more and more machine learning based programs are moving to cloud computing, SGX can be used in cloud-based Machine Learning applications to protect user data from malicious privileged programs.However, applications running in SGX suffer from several overheads, including frequent context switching, memory page encryption/decryption, and memory page swapping, which significantly degrade the execution efficiency. In this paper, we aim to i) comprehensively explore the execution of general AI applications running on SGX, ii) systematically characterize the data reuses at both page granularity and cacheline granularity, and iii) provide optimization insights for efficient deployment of machine learning based applications on SGX. To the best of our knowledge, our work is the first to study machine learning applications on SGX and explore the potential of data reuses to reduce the runtime overheads in SGX.","PeriodicalId":135930,"journal":{"name":"2021 IEEE International Conference on Networking, Architecture and Storage (NAS)","volume":"133 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115171338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Decoupling Control and Data Transmission in RDMA Enabled Cloud Data Centers 支持RDMA的云数据中心中的解耦控制与数据传输
Pub Date : 2021-10-01 DOI: 10.1109/nas51552.2021.9605415
Qingyue Liu, P. Varman
Advances in storage, processing, and networking hardware are changing the structure of distributed applications. RDMA networks provide multiple communication mechanisms that enable novel hybrid protocols specialized to different data transfer requirements. In this paper, we present a distributed communication scheme that separates control and data communication channels directly at the RNIC rather than the application level. We develop a new communication artifact, a remote random access buffer, to efficiently implement this separation. Data messages are sent silently to the receiver, which is informed of the location of the data by a subsequent control message. Experiments on an RDMA-enabled cluster with micro benchmarks and two distributed applications validate the performance benefits of our approach.
存储、处理和网络硬件方面的进步正在改变分布式应用程序的结构。RDMA网络提供了多种通信机制,使新的混合协议能够满足不同的数据传输需求。在本文中,我们提出了一种分布式通信方案,该方案直接在RNIC而不是应用程序级别分离控制和数据通信通道。我们开发了一个新的通信构件,一个远程随机访问缓冲区,以有效地实现这种分离。数据消息以静默方式发送给接收方,随后的控制消息通知接收方数据的位置。使用微基准测试和两个分布式应用程序在支持rdma的集群上进行的实验验证了我们的方法的性能优势。
{"title":"Decoupling Control and Data Transmission in RDMA Enabled Cloud Data Centers","authors":"Qingyue Liu, P. Varman","doi":"10.1109/nas51552.2021.9605415","DOIUrl":"https://doi.org/10.1109/nas51552.2021.9605415","url":null,"abstract":"Advances in storage, processing, and networking hardware are changing the structure of distributed applications. RDMA networks provide multiple communication mechanisms that enable novel hybrid protocols specialized to different data transfer requirements. In this paper, we present a distributed communication scheme that separates control and data communication channels directly at the RNIC rather than the application level. We develop a new communication artifact, a remote random access buffer, to efficiently implement this separation. Data messages are sent silently to the receiver, which is informed of the location of the data by a subsequent control message. Experiments on an RDMA-enabled cluster with micro benchmarks and two distributed applications validate the performance benefits of our approach.","PeriodicalId":135930,"journal":{"name":"2021 IEEE International Conference on Networking, Architecture and Storage (NAS)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127909522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
E2E Visual Analytics: Achieving >10X Edge/Cloud Optimizations E2E可视化分析:实现>10倍的边缘/云优化
Pub Date : 2021-10-01 DOI: 10.1109/nas51552.2021.9605404
Chaunté W. Lacewell, Nilesh A. Ahuja, Pablo Muñoz, Parual Datta, Ragaad Altarawneh, Vui Seng Chua, Nilesh Jain, Omesh Tickoo, R. Iyer
As visual analytics continues to rapidly grow, there is a critical need to improve the end-to-end efficiency of visual processing in edge/cloud systems. In this paper, we cover algorithms, systems and optimizations in three major areas for edge/cloud visual processing: (1) addressing storage and retrieval efficiency of visual data and meta-data by employing and optimizing visual data management systems, (2) addressing compute efficiency of visual analytics by taking advantage of co-optimization between the compression and analytics domains and (3) addressing networking (bandwidth) efficiency of visual data compression by tailoring it based on analytics tasks. We describe techniques in each of the above areas and measure its efficacy on state-of-the-art platforms (Intel Xeon), workloads and datasets. Our results show that we can achieve >10X improvements in each area based on novel algorithms, systems, and co-design optimizations. We also outline future research directions based on our findings which outline areas of further performance and efficiency advantages in end-to-end visual analytics.
随着视觉分析持续快速增长,迫切需要提高边缘/云系统中视觉处理的端到端效率。在本文中,我们涵盖了边缘/云视觉处理的三个主要领域的算法、系统和优化:(1)通过使用和优化可视化数据管理系统来解决可视化数据和元数据的存储和检索效率问题;(2)通过利用压缩和分析领域之间的协同优化来解决可视化分析的计算效率问题;(3)通过根据分析任务定制可视化数据压缩来解决可视化数据压缩的网络(带宽)效率问题。我们将描述上述每个领域的技术,并测量其在最先进的平台(Intel Xeon)、工作负载和数据集上的效率。我们的研究结果表明,基于新的算法、系统和协同设计优化,我们可以在每个领域实现>10倍的改进。我们还根据我们的发现概述了未来的研究方向,这些方向概述了端到端可视化分析中进一步性能和效率优势的领域。
{"title":"E2E Visual Analytics: Achieving >10X Edge/Cloud Optimizations","authors":"Chaunté W. Lacewell, Nilesh A. Ahuja, Pablo Muñoz, Parual Datta, Ragaad Altarawneh, Vui Seng Chua, Nilesh Jain, Omesh Tickoo, R. Iyer","doi":"10.1109/nas51552.2021.9605404","DOIUrl":"https://doi.org/10.1109/nas51552.2021.9605404","url":null,"abstract":"As visual analytics continues to rapidly grow, there is a critical need to improve the end-to-end efficiency of visual processing in edge/cloud systems. In this paper, we cover algorithms, systems and optimizations in three major areas for edge/cloud visual processing: (1) addressing storage and retrieval efficiency of visual data and meta-data by employing and optimizing visual data management systems, (2) addressing compute efficiency of visual analytics by taking advantage of co-optimization between the compression and analytics domains and (3) addressing networking (bandwidth) efficiency of visual data compression by tailoring it based on analytics tasks. We describe techniques in each of the above areas and measure its efficacy on state-of-the-art platforms (Intel Xeon), workloads and datasets. Our results show that we can achieve >10X improvements in each area based on novel algorithms, systems, and co-design optimizations. We also outline future research directions based on our findings which outline areas of further performance and efficiency advantages in end-to-end visual analytics.","PeriodicalId":135930,"journal":{"name":"2021 IEEE International Conference on Networking, Architecture and Storage (NAS)","volume":"26 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128226692","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On Adapting the Cache Block Size in SSD Caches 关于调整SSD缓存块大小的研究
Pub Date : 2021-10-01 DOI: 10.1109/nas51552.2021.9605462
Nikolaus Jeremic, Helge Parzyjegla, Gero Mühl
SSD-based block-level caches can notably increase the performance of HDD-based storage systems. However, this demands a sensible choice of the cache block size, which depends strongly on the workload characteristics. Many workloads will most likely favor either small or large cache blocks. Unfortunately, choosing the appropriate cache block size is difficult due to the diversity and dynamics of storage workloads. Thus, adapting the cache block size to the workload characteristics at run time has the potential to substantially improve the cache performance compared to using a fixed cache block size. However, changing the used cache block size for all cached data is very costly and neglects that distinct parts of the data may exhibit different access patterns, which favor distinct cache block sizes.In this paper, we experimentally study the performance impact of the cache block size and fine-grained adaptation, i.e., for individual parts of the data, between small and large cache blocks in write-back SSD caches. Based on our results, we make two major observations on the performance impact of the cache block size and its adaptation. First, using an inappropriate cache block size can reduce the overall throughput by up to 84% compared to using the most suitable cache block size. Second, fine-grained adaptation between small and large cache blocks is highly beneficial as it avoids such a performance deterioration, whereas it can increase the overall throughput by up to 126% in comparison to using the more suitable fixed cache block size.
基于ssd的块级缓存可以显著提高基于hdd的存储系统的性能。但是,这需要明智地选择缓存块大小,这在很大程度上取决于工作负载特征。许多工作负载很可能喜欢小缓存块或大缓存块。不幸的是,由于存储工作负载的多样性和动态性,选择适当的缓存块大小很困难。因此,与使用固定的缓存块大小相比,在运行时根据工作负载特征调整缓存块大小有可能大大提高缓存性能。但是,更改所有缓存数据的已用缓存块大小的成本非常高,并且忽略了数据的不同部分可能表现出不同的访问模式,这有利于不同的缓存块大小。在本文中,我们实验研究了缓存块大小和细粒度适应对性能的影响,即对于数据的单个部分,在回写SSD缓存中的小缓存块和大缓存块之间。基于我们的结果,我们对缓存块大小及其适应性对性能的影响进行了两个主要观察。首先,与使用最合适的缓存块大小相比,使用不合适的缓存块大小可以减少高达84%的总吞吐量。其次,小缓存块和大缓存块之间的细粒度适应非常有益,因为它避免了这种性能下降,而与使用更合适的固定缓存块大小相比,它可以将总吞吐量提高多达126%。
{"title":"On Adapting the Cache Block Size in SSD Caches","authors":"Nikolaus Jeremic, Helge Parzyjegla, Gero Mühl","doi":"10.1109/nas51552.2021.9605462","DOIUrl":"https://doi.org/10.1109/nas51552.2021.9605462","url":null,"abstract":"SSD-based block-level caches can notably increase the performance of HDD-based storage systems. However, this demands a sensible choice of the cache block size, which depends strongly on the workload characteristics. Many workloads will most likely favor either small or large cache blocks. Unfortunately, choosing the appropriate cache block size is difficult due to the diversity and dynamics of storage workloads. Thus, adapting the cache block size to the workload characteristics at run time has the potential to substantially improve the cache performance compared to using a fixed cache block size. However, changing the used cache block size for all cached data is very costly and neglects that distinct parts of the data may exhibit different access patterns, which favor distinct cache block sizes.In this paper, we experimentally study the performance impact of the cache block size and fine-grained adaptation, i.e., for individual parts of the data, between small and large cache blocks in write-back SSD caches. Based on our results, we make two major observations on the performance impact of the cache block size and its adaptation. First, using an inappropriate cache block size can reduce the overall throughput by up to 84% compared to using the most suitable cache block size. Second, fine-grained adaptation between small and large cache blocks is highly beneficial as it avoids such a performance deterioration, whereas it can increase the overall throughput by up to 126% in comparison to using the more suitable fixed cache block size.","PeriodicalId":135930,"journal":{"name":"2021 IEEE International Conference on Networking, Architecture and Storage (NAS)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131060944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Exploring Storage Device Characteristics of A RISC-V Little-core SoC RISC-V小核SoC的存储器件特性研究
Pub Date : 2021-10-01 DOI: 10.1109/nas51552.2021.9605430
Tao Lu
Low-power system-on-chips (SoCs) dominate the Internet of Things (IoT) ecosystem, which consists of billions of devices that can generate Zettabytes of data. SoC directly interacts with big data, but there is little research on its storage performance and power consumption characteristics, especially the lack of quantitative evaluation. In this paper, we study the storage characteristics of a low-power RISC-V SoC FPGA. Specifically, we deploy a PCIe SSD to study the performance of storage devices under little cores. We quantitatively evaluate device bandwidth, IOPS throughput, and power consumption. In addition, we compare the same device on the low-power RISC-V SoC and a high-performance x86 server to observe the similarities and differences of the storage device behavior on different computing platforms.
低功耗的片上系统(soc)在物联网(IoT)生态系统中占据主导地位,物联网由数十亿台设备组成,这些设备可以产生zb级的数据。SoC直接与大数据交互,但对其存储性能和功耗特性的研究很少,特别是缺乏定量评价。本文研究了一种低功耗RISC-V SoC FPGA的存储特性。具体来说,我们部署了一个PCIe固态硬盘来研究小核下存储设备的性能。我们定量评估设备带宽、IOPS吞吐量和功耗。此外,我们将同一设备在低功耗RISC-V SoC和高性能x86服务器上进行比较,观察不同计算平台上存储设备行为的异同。
{"title":"Exploring Storage Device Characteristics of A RISC-V Little-core SoC","authors":"Tao Lu","doi":"10.1109/nas51552.2021.9605430","DOIUrl":"https://doi.org/10.1109/nas51552.2021.9605430","url":null,"abstract":"Low-power system-on-chips (SoCs) dominate the Internet of Things (IoT) ecosystem, which consists of billions of devices that can generate Zettabytes of data. SoC directly interacts with big data, but there is little research on its storage performance and power consumption characteristics, especially the lack of quantitative evaluation. In this paper, we study the storage characteristics of a low-power RISC-V SoC FPGA. Specifically, we deploy a PCIe SSD to study the performance of storage devices under little cores. We quantitatively evaluate device bandwidth, IOPS throughput, and power consumption. In addition, we compare the same device on the low-power RISC-V SoC and a high-performance x86 server to observe the similarities and differences of the storage device behavior on different computing platforms.","PeriodicalId":135930,"journal":{"name":"2021 IEEE International Conference on Networking, Architecture and Storage (NAS)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131409951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Balancing Latency and Quality in Web Search 平衡网络搜索的延迟和质量
Pub Date : 2021-10-01 DOI: 10.1109/nas51552.2021.9605375
Liang Zhou, K. Ramakrishnan
Selecting the right time budget for a search query is challenging because a proper balance between the search latency, quality and efficiency has to be maintained. State-of-the-art approaches leverage a centralized sample index at the aggregator to select the Index Serving Nodes (ISNs) to maintain quality and responsiveness. In this paper, we propose Cottage, a coordinated framework between the aggregator and ISNs for latency and quality optimization in web search. Cottage has two separate neural network models at each ISN to predict the quality contribution and latency, respectively. Then, these prediction results are sent back to the aggregator for latency and quality optimizations. The key task is integration of the predictions at the aggregator in determining an optimal dynamic time budget for identifying slow and low quality ISNs to improve latency and search efficiency. Our experiments on the Solr search engine prove that Cottage can reduce the average query latency by 54% and achieve a good P@10 search quality of 0.947.
为搜索查询选择合适的时间预算是一项挑战,因为必须在搜索延迟、质量和效率之间保持适当的平衡。最先进的方法利用聚合器上的集中式样本索引来选择索引服务节点(index service Nodes, isn),以保持质量和响应能力。在本文中,我们提出了Cottage,这是一个聚合器和ISNs之间的协调框架,用于网络搜索中的延迟和质量优化。Cottage在每个ISN上都有两个独立的神经网络模型,分别预测质量贡献和延迟。然后,这些预测结果被发送回聚合器进行延迟和质量优化。关键任务是在聚合器中集成预测,确定最优动态时间预算,用于识别缓慢和低质量的isn,以提高延迟和搜索效率。我们在Solr搜索引擎上的实验证明,Cottage可以将平均查询延迟减少54%,并获得0.947的良好P@10搜索质量。
{"title":"Balancing Latency and Quality in Web Search","authors":"Liang Zhou, K. Ramakrishnan","doi":"10.1109/nas51552.2021.9605375","DOIUrl":"https://doi.org/10.1109/nas51552.2021.9605375","url":null,"abstract":"Selecting the right time budget for a search query is challenging because a proper balance between the search latency, quality and efficiency has to be maintained. State-of-the-art approaches leverage a centralized sample index at the aggregator to select the Index Serving Nodes (ISNs) to maintain quality and responsiveness. In this paper, we propose Cottage, a coordinated framework between the aggregator and ISNs for latency and quality optimization in web search. Cottage has two separate neural network models at each ISN to predict the quality contribution and latency, respectively. Then, these prediction results are sent back to the aggregator for latency and quality optimizations. The key task is integration of the predictions at the aggregator in determining an optimal dynamic time budget for identifying slow and low quality ISNs to improve latency and search efficiency. Our experiments on the Solr search engine prove that Cottage can reduce the average query latency by 54% and achieve a good P@10 search quality of 0.947.","PeriodicalId":135930,"journal":{"name":"2021 IEEE International Conference on Networking, Architecture and Storage (NAS)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120894459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Flow Scheduling in a Heterogeneous NFV Environment using Reinforcement Learning 基于强化学习的异构NFV环境流调度
Pub Date : 2021-10-01 DOI: 10.1109/nas51552.2021.9605395
Chun Jen Lin, Yan Luo, Liang-Min Wang, Li-De Chen
Network function virtualization (NFV) allows net-work functions executed on general-purpose servers or virtual machines (VMs) instead of proprietary hardware, greatly improving the flexibility and scalability of network services. Recent trends in using programmable accelerators to speed up NFV performance introduce challenges in flow scheduling in a dynamic NFV environment. Reinforcement learning (RL) trains machine learning models for decision making to maximize returns in uncertain environments such as NFV. In this paper, we study the allocation of heterogeneous processors (CPUs and FPGAs) to minimize the delays of flows in the system. We conduct extensive simulations to evaluate the performance of reinforcement learning based scheduling algorithms such as Advantage Actor Critic (A2C), Trust Region Policy Optimization (TRPO) and Proximal Policy Optimization (PPO), and compare with greedy policies. The results show that RL based schedulers can effectively learn from past experiences and converge to the optimal greedy policy. We also analyze in-depth how the policies lead to different processor utilization and flow processing time, and provide insights into these policies.
网络功能虚拟化(NFV)允许在通用服务器或虚拟机上执行网络功能,而不是在专用硬件上执行,从而大大提高了网络服务的灵活性和可伸缩性。最近使用可编程加速器加速NFV性能的趋势给动态NFV环境中的流量调度带来了挑战。强化学习(RL)训练机器学习模型,用于决策制定,以在NFV等不确定环境中实现回报最大化。在本文中,我们研究了异构处理器(cpu和fpga)的分配,以最小化系统中的流延迟。我们进行了大量的模拟,以评估基于强化学习的调度算法的性能,如优势参与者批评家(A2C),信任区域策略优化(TRPO)和近端策略优化(PPO),并与贪婪策略进行比较。结果表明,基于强化学习的调度程序可以有效地从过去的经验中学习,并收敛到最优贪心策略。我们还深入分析了策略如何导致不同的处理器利用率和流处理时间,并提供了对这些策略的见解。
{"title":"Flow Scheduling in a Heterogeneous NFV Environment using Reinforcement Learning","authors":"Chun Jen Lin, Yan Luo, Liang-Min Wang, Li-De Chen","doi":"10.1109/nas51552.2021.9605395","DOIUrl":"https://doi.org/10.1109/nas51552.2021.9605395","url":null,"abstract":"Network function virtualization (NFV) allows net-work functions executed on general-purpose servers or virtual machines (VMs) instead of proprietary hardware, greatly improving the flexibility and scalability of network services. Recent trends in using programmable accelerators to speed up NFV performance introduce challenges in flow scheduling in a dynamic NFV environment. Reinforcement learning (RL) trains machine learning models for decision making to maximize returns in uncertain environments such as NFV. In this paper, we study the allocation of heterogeneous processors (CPUs and FPGAs) to minimize the delays of flows in the system. We conduct extensive simulations to evaluate the performance of reinforcement learning based scheduling algorithms such as Advantage Actor Critic (A2C), Trust Region Policy Optimization (TRPO) and Proximal Policy Optimization (PPO), and compare with greedy policies. The results show that RL based schedulers can effectively learn from past experiences and converge to the optimal greedy policy. We also analyze in-depth how the policies lead to different processor utilization and flow processing time, and provide insights into these policies.","PeriodicalId":135930,"journal":{"name":"2021 IEEE International Conference on Networking, Architecture and Storage (NAS)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127829280","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 IEEE International Conference on Networking, Architecture and Storage (NAS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1