首页 > 最新文献

arXiv - CS - Operating Systems最新文献

英文 中文
Configuration Validation with Large Language Models 大型语言模型的配置验证
Pub Date : 2023-10-15 DOI: arxiv-2310.09690
Xinyu Lian, Yinfang Chen, Runxiang Cheng, Jie Huang, Parth Thakkar, Tianyin Xu
Misconfigurations are the major causes of software failures. Existingconfiguration validation techniques rely on manually written rules or testcases, which are expensive to implement and maintain, and are hard to becomprehensive. Leveraging machine learning (ML) and natural language processing(NLP) for configuration validation is considered a promising direction, but hasbeen facing challenges such as the need of not only large-scale configurationdata, but also system-specific features and models which are hard togeneralize. Recent advances in Large Language Models (LLMs) show the promisesto address some of the long-lasting limitations of ML/NLP-based configurationvalidation techniques. In this paper, we present an exploratory analysis on thefeasibility and effectiveness of using LLMs like GPT and Codex forconfiguration validation. Specifically, we take a first step to empiricallyevaluate LLMs as configuration validators without additional fine-tuning orcode generation. We develop a generic LLM-based validation framework, namedCiri, which integrates different LLMs. Ciri devises effective promptengineering with few-shot learning based on both valid configuration andmisconfiguration data. Ciri also validates and aggregates the outputs of LLMsto generate validation results, coping with known hallucination andnondeterminism of LLMs. We evaluate the validation effectiveness of Ciri onfive popular LLMs using configuration data of six mature, widely deployedopen-source systems. Our analysis (1) confirms the potential of using LLMs forconfiguration validation, (2) understands the design space of LLMbasedvalidators like Ciri, especially in terms of prompt engineering with few-shotlearning, and (3) reveals open challenges such as ineffectiveness in detectingcertain types of misconfigurations and biases to popular configurationparameters.
错误配置是导致软件故障的主要原因。现有的配置验证技术依赖于手工编写的规则或测试用例,这在实现和维护上是昂贵的,并且很难理解。利用机器学习(ML)和自然语言处理(NLP)进行配置验证被认为是一个有前途的方向,但一直面临着挑战,例如不仅需要大规模配置数据,而且还需要难以概括的系统特定功能和模型。大型语言模型(llm)的最新进展表明,它有望解决基于ML/ nlp的配置验证技术的一些长期限制。在本文中,我们对使用像GPT和Codex这样的llm进行配置验证的可行性和有效性进行了探索性分析。具体地说,我们采取了第一步,在没有额外的微调或代码生成的情况下,将llm作为配置验证器进行经验评估。我们开发了一个通用的基于llm的验证框架,名为ciri,它集成了不同的llm。Ciri设计了有效的提示工程,基于有效配置和错误配置数据的少量学习。Ciri还对llm的输出进行验证和聚合以生成验证结果,以应对已知的llm的幻觉和不确定性。我们使用六个成熟的、广泛部署的开源系统的配置数据,评估了Ciri在五个流行的llm上的验证有效性。我们的分析(1)证实了使用llm进行配置验证的潜力,(2)了解了基于llm的验证器(如Ciri)的设计空间,特别是在使用少量学习的快速工程方面,以及(3)揭示了开放的挑战,例如检测某些类型的错误配置的有效性和对流行配置参数的偏差。
{"title":"Configuration Validation with Large Language Models","authors":"Xinyu Lian, Yinfang Chen, Runxiang Cheng, Jie Huang, Parth Thakkar, Tianyin Xu","doi":"arxiv-2310.09690","DOIUrl":"https://doi.org/arxiv-2310.09690","url":null,"abstract":"Misconfigurations are the major causes of software failures. Existing\u0000configuration validation techniques rely on manually written rules or test\u0000cases, which are expensive to implement and maintain, and are hard to be\u0000comprehensive. Leveraging machine learning (ML) and natural language processing\u0000(NLP) for configuration validation is considered a promising direction, but has\u0000been facing challenges such as the need of not only large-scale configuration\u0000data, but also system-specific features and models which are hard to\u0000generalize. Recent advances in Large Language Models (LLMs) show the promises\u0000to address some of the long-lasting limitations of ML/NLP-based configuration\u0000validation techniques. In this paper, we present an exploratory analysis on the\u0000feasibility and effectiveness of using LLMs like GPT and Codex for\u0000configuration validation. Specifically, we take a first step to empirically\u0000evaluate LLMs as configuration validators without additional fine-tuning or\u0000code generation. We develop a generic LLM-based validation framework, named\u0000Ciri, which integrates different LLMs. Ciri devises effective prompt\u0000engineering with few-shot learning based on both valid configuration and\u0000misconfiguration data. Ciri also validates and aggregates the outputs of LLMs\u0000to generate validation results, coping with known hallucination and\u0000nondeterminism of LLMs. We evaluate the validation effectiveness of Ciri on\u0000five popular LLMs using configuration data of six mature, widely deployed\u0000open-source systems. Our analysis (1) confirms the potential of using LLMs for\u0000configuration validation, (2) understands the design space of LLMbased\u0000validators like Ciri, especially in terms of prompt engineering with few-shot\u0000learning, and (3) reveals open challenges such as ineffectiveness in detecting\u0000certain types of misconfigurations and biases to popular configuration\u0000parameters.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"56 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138522171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Taking the Shortcut: Actively Incorporating the Virtual Memory Index of the OS to Hardware-Accelerate Database Indexing 走捷径:积极将操作系统的虚拟内存索引整合到硬件加速数据库索引中
Pub Date : 2023-10-13 DOI: arxiv-2310.09124
Felix Schuhknecht
Index structures often materialize one or multiple levels of explicitindirections (aka pointers) to allow for a quick traversal to the data ofinterest. Unfortunately, dereferencing a pointer to go from one level to theother is costly since additionally to following the address, it involves twoaddress translations from virtual memory to physical memory under the hood. Inthe worst case, such an address translation is resolved by an index accessitself, namely by a lookup into the page table, a central hardware-acceleratedindex structure of the OS. However, if the page table is anyways constantlyqueried, it raises the question whether we can actively incorporate it into ourdatabase indexes and make it work for us. Precisely, instead of materializingindirections in form of pointers, we propose to express these indirectionsdirectly in the page table wherever possible. By introducing such shortcuts, we(a) effectively reduce the height of traversal during lookups and (b) exploitthe hardware-acceleration of lookups in the page table. In this work, weanalyze the strengths and considerations of this approach and showcase itseffectiveness at the case of the real-world indexing scheme extendible hashing.
索引结构通常具体化一个或多个层次的显式间接指示(又名指针),以允许快速遍历感兴趣的数据。不幸的是,将指针从一个级别解引用到另一个级别是非常昂贵的,因为除了跟踪地址之外,它还涉及从虚拟内存到物理内存的两次地址转换。在最坏的情况下,这样的地址转换由索引访问本身来解决,即通过对页表的查找来解决,页表是操作系统的一个中央硬件加速索引结构。然而,如果页表总是被查询,那么我们是否可以积极地将它合并到我们的数据库索引中,并使其为我们工作。确切地说,我们建议尽可能在页表中直接表达这些间接,而不是以指针的形式具体化间接。通过引入这样的快捷方式,我们(a)有效地降低了查找过程中的遍历高度,(b)利用了页表中查找的硬件加速。在这项工作中,我们分析了这种方法的优势和注意事项,并展示了它在现实世界索引方案可扩展散列中的有效性。
{"title":"Taking the Shortcut: Actively Incorporating the Virtual Memory Index of the OS to Hardware-Accelerate Database Indexing","authors":"Felix Schuhknecht","doi":"arxiv-2310.09124","DOIUrl":"https://doi.org/arxiv-2310.09124","url":null,"abstract":"Index structures often materialize one or multiple levels of explicit\u0000indirections (aka pointers) to allow for a quick traversal to the data of\u0000interest. Unfortunately, dereferencing a pointer to go from one level to the\u0000other is costly since additionally to following the address, it involves two\u0000address translations from virtual memory to physical memory under the hood. In\u0000the worst case, such an address translation is resolved by an index access\u0000itself, namely by a lookup into the page table, a central hardware-accelerated\u0000index structure of the OS. However, if the page table is anyways constantly\u0000queried, it raises the question whether we can actively incorporate it into our\u0000database indexes and make it work for us. Precisely, instead of materializing\u0000indirections in form of pointers, we propose to express these indirections\u0000directly in the page table wherever possible. By introducing such shortcuts, we\u0000(a) effectively reduce the height of traversal during lookups and (b) exploit\u0000the hardware-acceleration of lookups in the page table. In this work, we\u0000analyze the strengths and considerations of this approach and showcase its\u0000effectiveness at the case of the real-world indexing scheme extendible hashing.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"24 4","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138522170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards a debuggable kernel design 迈向可调试的内核设计
Pub Date : 2023-10-09 DOI: arxiv-2310.05399
Chandrika Parimoo, Ashish Gupta
This paper describes what it means for a kernel to be debuggable and proposesa kernel design with debuggability in mind. We evaluate the proposed kerneldesign by comparing the iterations required in cyclic debugging for differentclasses of bugs in a vanilla monolithic kernel to a variant enhanced with ourdesign rules for debuggability. We discuss the trade offs involved in designinga debuggable kernel.
本文描述了内核可调试的含义,并提出了一种考虑可调试性的内核设计。我们通过比较在一个普通的单片内核中对不同类型的bug进行循环调试所需的迭代,来评估所提出的内核设计。我们讨论了在设计可调试内核时所涉及的权衡。
{"title":"Towards a debuggable kernel design","authors":"Chandrika Parimoo, Ashish Gupta","doi":"arxiv-2310.05399","DOIUrl":"https://doi.org/arxiv-2310.05399","url":null,"abstract":"This paper describes what it means for a kernel to be debuggable and proposes\u0000a kernel design with debuggability in mind. We evaluate the proposed kernel\u0000design by comparing the iterations required in cyclic debugging for different\u0000classes of bugs in a vanilla monolithic kernel to a variant enhanced with our\u0000design rules for debuggability. We discuss the trade offs involved in designing\u0000a debuggable kernel.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"30 7","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138522100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prompt-to-OS (P2OS): Revolutionizing Operating Systems and Human-Computer Interaction with Integrated AI Generative Models 即时操作系统(P2OS):革命性的操作系统和集成人工智能生成模型的人机交互
Pub Date : 2023-10-07 DOI: arxiv-2310.04875
Gabriele Tolomei, Cesare Campagnano, Fabrizio Silvestri, Giovanni Trappolini
In this paper, we present a groundbreaking paradigm for human-computerinteraction that revolutionizes the traditional notion of an operating system. Within this innovative framework, user requests issued to the machine arehandled by an interconnected ecosystem of generative AI models that seamlesslyintegrate with or even replace traditional software applications. At the coreof this paradigm shift are large generative models, such as language anddiffusion models, which serve as the central interface between users andcomputers. This pioneering approach leverages the abilities of advancedlanguage models, empowering users to engage in natural language conversationswith their computing devices. Users can articulate their intentions, tasks, andinquiries directly to the system, eliminating the need for explicit commands orcomplex navigation. The language model comprehends and interprets the user'sprompts, generating and displaying contextual and meaningful responses thatfacilitate seamless and intuitive interactions. This paradigm shift not only streamlines user interactions but also opens upnew possibilities for personalized experiences. Generative models can adapt toindividual preferences, learning from user input and continuously improvingtheir understanding and response generation. Furthermore, it enables enhancedaccessibility, as users can interact with the system using speech or text,accommodating diverse communication preferences. However, this visionary concept raises significant challenges, includingprivacy, security, trustability, and the ethical use of generative models.Robust safeguards must be in place to protect user data and prevent potentialmisuse or manipulation of the language model. While the full realization of this paradigm is still far from being achieved,this paper serves as a starting point for envisioning this transformativepotential.
在本文中,我们提出了一个开创性的人机交互范例,它彻底改变了操作系统的传统概念。在这个创新的框架中,用户向机器发出的请求由一个由生成式人工智能模型组成的互联生态系统处理,这些模型与传统的软件应用程序无缝集成,甚至取代传统的软件应用程序。这种范式转变的核心是大型生成模型,如语言和扩散模型,它们作为用户和计算机之间的中心接口。这种开创性的方法利用了高级语言模型的能力,使用户能够与他们的计算设备进行自然语言对话。用户可以直接向系统表达他们的意图、任务和查询,从而消除了明确命令或复杂导航的需要。语言模型理解和解释用户的提示,生成和显示上下文和有意义的响应,促进无缝和直观的交互。这种模式的转变不仅简化了用户交互,而且为个性化体验开辟了新的可能性。生成模型可以适应个人偏好,从用户输入中学习,并不断提高他们的理解和响应生成。此外,它还增强了可访问性,因为用户可以使用语音或文本与系统交互,以适应不同的通信偏好。然而,这个有远见的概念提出了重大挑战,包括隐私、安全、可信赖性和生成模型的道德使用。必须有强大的保护措施来保护用户数据并防止对语言模型的潜在滥用或操纵。虽然这种范式的完全实现仍远未实现,但本文可以作为设想这种变革潜力的起点。
{"title":"Prompt-to-OS (P2OS): Revolutionizing Operating Systems and Human-Computer Interaction with Integrated AI Generative Models","authors":"Gabriele Tolomei, Cesare Campagnano, Fabrizio Silvestri, Giovanni Trappolini","doi":"arxiv-2310.04875","DOIUrl":"https://doi.org/arxiv-2310.04875","url":null,"abstract":"In this paper, we present a groundbreaking paradigm for human-computer\u0000interaction that revolutionizes the traditional notion of an operating system. Within this innovative framework, user requests issued to the machine are\u0000handled by an interconnected ecosystem of generative AI models that seamlessly\u0000integrate with or even replace traditional software applications. At the core\u0000of this paradigm shift are large generative models, such as language and\u0000diffusion models, which serve as the central interface between users and\u0000computers. This pioneering approach leverages the abilities of advanced\u0000language models, empowering users to engage in natural language conversations\u0000with their computing devices. Users can articulate their intentions, tasks, and\u0000inquiries directly to the system, eliminating the need for explicit commands or\u0000complex navigation. The language model comprehends and interprets the user's\u0000prompts, generating and displaying contextual and meaningful responses that\u0000facilitate seamless and intuitive interactions. This paradigm shift not only streamlines user interactions but also opens up\u0000new possibilities for personalized experiences. Generative models can adapt to\u0000individual preferences, learning from user input and continuously improving\u0000their understanding and response generation. Furthermore, it enables enhanced\u0000accessibility, as users can interact with the system using speech or text,\u0000accommodating diverse communication preferences. However, this visionary concept raises significant challenges, including\u0000privacy, security, trustability, and the ethical use of generative models.\u0000Robust safeguards must be in place to protect user data and prevent potential\u0000misuse or manipulation of the language model. While the full realization of this paradigm is still far from being achieved,\u0000this paper serves as a starting point for envisioning this transformative\u0000potential.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"39 12","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138522091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Victima: Drastically Increasing Address Translation Reach by Leveraging Underutilized Cache Resources 受害者:通过利用未充分利用的缓存资源大幅增加地址转换范围
Pub Date : 2023-10-06 DOI: arxiv-2310.04158
Konstantinos Kanellopoulos, Hong Chul Nam, F. Nisa Bostanci, Rahul Bera, Mohammad Sadrosadati, Rakesh Kumar, Davide-Basilio Bartolini, Onur Mutlu
Address translation is a performance bottleneck in data-intensive workloadsdue to large datasets and irregular access patterns that lead to frequenthigh-latency page table walks (PTWs). PTWs can be reduced by using (i) largehardware TLBs or (ii) large software-managed TLBs. Unfortunately, bothsolutions have significant drawbacks: increased access latency, power and area(for hardware TLBs), and costly memory accesses, the need for large contiguousmemory blocks, and complex OS modifications (for software-managed TLBs). Wepresent Victima, a new software-transparent mechanism that drasticallyincreases the translation reach of the processor by leveraging theunderutilized resources of the cache hierarchy. The key idea of Victima is torepurpose L2 cache blocks to store clusters of TLB entries, thereby providingan additional low-latency and high-capacity component that backs up thelast-level TLB and thus reduces PTWs. Victima has two main components. First, aPTW cost predictor (PTW-CP) identifies costly-to-translate addresses based onthe frequency and cost of the PTWs they lead to. Second, a TLB-aware cachereplacement policy prioritizes keeping TLB entries in the cache hierarchy byconsidering (i) the translation pressure (e.g., last-level TLB miss rate) and(ii) the reuse characteristics of the TLB entries. Our evaluation results showthat in native (virtualized) execution environments Victima improves averageend-to-end application performance by 7.4% (28.7%) over the baseline four-levelradix-tree-based page table design and by 6.2% (20.1%) over a state-of-the-artsoftware-managed TLB, across 11 diverse data-intensive workloads. Victima (i)is effective in both native and virtualized environments, (ii) is completelytransparent to application and system software, and (iii) incurs very smallarea and power overheads on a modern high-end CPU.
地址转换是数据密集型工作负载中的性能瓶颈,因为大型数据集和不规则的访问模式会导致频繁的高延迟页表遍历(PTWs)。可以通过使用(i)大型硬件tlb或(ii)大型软件管理的tlb来减少ptw。不幸的是,这两种解决方案都有明显的缺点:增加访问延迟、功率和面积(对于硬件tlb)、昂贵的内存访问、需要大的连续内存块以及复杂的操作系统修改(对于软件管理的tlb)。我们提出了受害者,一个新的软件透明机制,通过利用缓存层次结构中未充分利用的资源,极大地增加了处理器的翻译范围。受害的关键思想是重新利用L2缓存块来存储TLB条目的集群,从而提供一个额外的低延迟和高容量组件来备份最后一级TLB,从而减少ptw。受害者有两个主要组成部分。首先,aPTW成本预测器(PTW-CP)根据它们所导致的ptw的频率和成本来识别转换成本高的地址。其次,TLB感知缓存替换策略通过考虑(i)转换压力(例如,最后一层TLB缺失率)和(ii) TLB项的重用特征来优先保留TLB项在缓存层次结构中。我们的评估结果表明,在原生(虚拟化)执行环境中,在11种不同的数据密集型工作负载中,受害服务器的端到端应用程序性能比基于基数树的基线四层页表设计提高了7.4%(28.7%),比最先进的软件管理的TLB提高了6.2%(20.1%)。受害者(i)在本地和虚拟环境中都有效,(ii)对应用程序和系统软件完全透明,(iii)在现代高端CPU上占用非常小的面积和功率开销。
{"title":"Victima: Drastically Increasing Address Translation Reach by Leveraging Underutilized Cache Resources","authors":"Konstantinos Kanellopoulos, Hong Chul Nam, F. Nisa Bostanci, Rahul Bera, Mohammad Sadrosadati, Rakesh Kumar, Davide-Basilio Bartolini, Onur Mutlu","doi":"arxiv-2310.04158","DOIUrl":"https://doi.org/arxiv-2310.04158","url":null,"abstract":"Address translation is a performance bottleneck in data-intensive workloads\u0000due to large datasets and irregular access patterns that lead to frequent\u0000high-latency page table walks (PTWs). PTWs can be reduced by using (i) large\u0000hardware TLBs or (ii) large software-managed TLBs. Unfortunately, both\u0000solutions have significant drawbacks: increased access latency, power and area\u0000(for hardware TLBs), and costly memory accesses, the need for large contiguous\u0000memory blocks, and complex OS modifications (for software-managed TLBs). We\u0000present Victima, a new software-transparent mechanism that drastically\u0000increases the translation reach of the processor by leveraging the\u0000underutilized resources of the cache hierarchy. The key idea of Victima is to\u0000repurpose L2 cache blocks to store clusters of TLB entries, thereby providing\u0000an additional low-latency and high-capacity component that backs up the\u0000last-level TLB and thus reduces PTWs. Victima has two main components. First, a\u0000PTW cost predictor (PTW-CP) identifies costly-to-translate addresses based on\u0000the frequency and cost of the PTWs they lead to. Second, a TLB-aware cache\u0000replacement policy prioritizes keeping TLB entries in the cache hierarchy by\u0000considering (i) the translation pressure (e.g., last-level TLB miss rate) and\u0000(ii) the reuse characteristics of the TLB entries. Our evaluation results show\u0000that in native (virtualized) execution environments Victima improves average\u0000end-to-end application performance by 7.4% (28.7%) over the baseline four-level\u0000radix-tree-based page table design and by 6.2% (20.1%) over a state-of-the-art\u0000software-managed TLB, across 11 diverse data-intensive workloads. Victima (i)\u0000is effective in both native and virtualized environments, (ii) is completely\u0000transparent to application and system software, and (iii) incurs very small\u0000area and power overheads on a modern high-end CPU.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"24 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138522169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Motivating Next-Generation OS Physical Memory Management for Terabyte-Scale NVMMs 推动下一代操作系统物理内存管理的tb级nvmm
Pub Date : 2023-10-05 DOI: arxiv-2310.03370
Shivank Garg, Aravinda Prasad, Debadatta Mishra, Sreenivas Subramoney
Software managed byte-addressable hybrid memory systems consisting of DRAMsand NVMMs offer a lot of flexibility to design efficient large scale dataprocessing applications. Operating systems (OS) play an important role inenabling the applications to realize the integrated benefits of DRAMs' lowaccess latency and NVMMs' large capacity along with its persistentcharacteristics. In this paper, we comprehensively analyze the performance ofconventional OS physical memory management subsystems that were designed onlybased on the DRAM memory characteristics in the context of modern hybridbyte-addressable memory systems. To study the impact of high access latency and large capacity of NVMMs onphysical memory management, we perform an extensive evaluation on Linux withIntel's Optane NVMM. We observe that the core memory management functionalitiessuch as page allocation are negatively impacted by high NVMM media latency,while functionalities such as conventional fragmentation management arerendered inadequate. We also demonstrate that certain traditional memorymanagement functionalities are affected by neither aspects of modern NVMMs. Weconclusively motivate the need to overhaul fundamental aspects of traditionalOS physical memory management in order to fully exploit terabyte-scale NVMMs.
软件管理的由dram和nvmm组成的可寻址字节混合存储系统为设计高效的大规模数据处理应用程序提供了很大的灵活性。操作系统(OS)在使应用程序实现dram的低访问延迟和nvmm的大容量及其持久特性的综合优势方面发挥着重要作用。本文在现代混合字节可寻址存储系统的背景下,全面分析了仅基于DRAM存储特性设计的传统操作系统物理内存管理子系统的性能。为了研究高访问延迟和大容量NVMM对物理内存管理的影响,我们在Linux上使用intel的Optane NVMM进行了广泛的评估。我们观察到,核心内存管理功能(如页面分配)受到高NVMM媒体延迟的负面影响,而传统碎片管理等功能则呈现不足。我们还证明了某些传统的内存管理功能不受现代nvmm的任何方面的影响。为了充分利用tb级的nvmm,我们最终激发了对传统os物理内存管理的基本方面进行彻底检修的需求。
{"title":"Motivating Next-Generation OS Physical Memory Management for Terabyte-Scale NVMMs","authors":"Shivank Garg, Aravinda Prasad, Debadatta Mishra, Sreenivas Subramoney","doi":"arxiv-2310.03370","DOIUrl":"https://doi.org/arxiv-2310.03370","url":null,"abstract":"Software managed byte-addressable hybrid memory systems consisting of DRAMs\u0000and NVMMs offer a lot of flexibility to design efficient large scale data\u0000processing applications. Operating systems (OS) play an important role in\u0000enabling the applications to realize the integrated benefits of DRAMs' low\u0000access latency and NVMMs' large capacity along with its persistent\u0000characteristics. In this paper, we comprehensively analyze the performance of\u0000conventional OS physical memory management subsystems that were designed only\u0000based on the DRAM memory characteristics in the context of modern hybrid\u0000byte-addressable memory systems. To study the impact of high access latency and large capacity of NVMMs on\u0000physical memory management, we perform an extensive evaluation on Linux with\u0000Intel's Optane NVMM. We observe that the core memory management functionalities\u0000such as page allocation are negatively impacted by high NVMM media latency,\u0000while functionalities such as conventional fragmentation management are\u0000rendered inadequate. We also demonstrate that certain traditional memory\u0000management functionalities are affected by neither aspects of modern NVMMs. We\u0000conclusively motivate the need to overhaul fundamental aspects of traditional\u0000OS physical memory management in order to fully exploit terabyte-scale NVMMs.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"26 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138542488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Co-Optimizing Cache Partitioning and Multi-Core Task Scheduling: Exploit Cache Sensitivity or Not? 协同优化缓存分区和多核任务调度:利用缓存敏感性与否?
Pub Date : 2023-10-04 DOI: arxiv-2310.02959
Binqi Sun, Debayan Roy, Tomasz Kloda, Andrea Bastoni, Rodolfo Pellizzoni, Marco Caccamo
Cache partitioning techniques have been successfully adopted to mitigateinterference among concurrently executing real-time tasks on multi-coreprocessors. Considering that the execution time of a cache-sensitive taskstrongly depends on the cache available for it to use, co-optimizing cachepartitioning and task allocation improves the system's schedulability. In thispaper, we propose a hybrid multi-layer design space exploration technique tosolve this multi-resource management problem. We explore the interplay betweencache partitioning and schedulability by systematically interleaving threeoptimization layers, viz., (i) in the outer layer, we perform a breadth-firstsearch combined with proactive pruning for cache partitioning; (ii) in themiddle layer, we exploit a first-fit heuristic for allocating tasks to cores;and (iii) in the inner layer, we use the well-known recurrence relation for theschedulability analysis of non-preemptive fixed-priority (NP-FP) tasks in auniprocessor setting. Although our focus is on NP-FP scheduling, we evaluatethe flexibility of our framework in supporting different scheduling policies(NP-EDF, P-EDF) by plugging in appropriate analysis methods in the inner layer.Experiments show that, compared to the state-of-the-art techniques, theproposed framework can improve the real-time schedulability of NP-FP task setsby an average of 15.2% with a maximum improvement of 233.6% (when tasks arehighly cache-sensitive) and a minimum of 1.6% (when cache sensitivity is low).For such task sets, we found that clustering similar-period (or mutuallycompatible) tasks often leads to higher schedulability (on average 7.6%) thanclustering by cache sensitivity. In our evaluation, the framework also achievesgood results for preemptive and dynamic-priority scheduling policies.
缓存分区技术已被成功地用于减轻多核处理器上并发执行实时任务之间的干扰。考虑到对缓存敏感的任务的执行时间很大程度上依赖于可供其使用的缓存,因此协同优化缓存分区和任务分配可以提高系统的可调度性。本文提出了一种混合多层设计空间探索技术来解决这一多资源管理问题。我们通过系统地交错三个优化层来探索缓存分区和可调度性之间的相互作用,即:(i)在外层,我们执行宽度优先搜索并结合主动修剪缓存分区;(ii)在中间层,我们利用第一拟合启发法将任务分配给核心;(iii)在内层,我们使用众所周知的递归关系来分析单处理器设置下非抢占式固定优先级(NP-FP)任务的可调度性。虽然我们的重点是NP-FP调度,但我们通过在内层插入适当的分析方法来评估我们的框架在支持不同调度策略(NP-EDF, P-EDF)方面的灵活性。实验表明,与最先进的技术相比,所提出的框架可以提高NP-FP任务集的实时可调度性,平均提高15.2%,最大提高233.6%(当任务对缓存高度敏感时),最小提高1.6%(当缓存灵敏度较低时)。对于这样的任务集,我们发现聚类相似周期(或相互兼容)的任务通常比通过缓存灵敏度聚类获得更高的可调度性(平均7.6%)。在我们的评估中,该框架在抢占和动态优先级调度策略方面也取得了良好的效果。
{"title":"Co-Optimizing Cache Partitioning and Multi-Core Task Scheduling: Exploit Cache Sensitivity or Not?","authors":"Binqi Sun, Debayan Roy, Tomasz Kloda, Andrea Bastoni, Rodolfo Pellizzoni, Marco Caccamo","doi":"arxiv-2310.02959","DOIUrl":"https://doi.org/arxiv-2310.02959","url":null,"abstract":"Cache partitioning techniques have been successfully adopted to mitigate\u0000interference among concurrently executing real-time tasks on multi-core\u0000processors. Considering that the execution time of a cache-sensitive task\u0000strongly depends on the cache available for it to use, co-optimizing cache\u0000partitioning and task allocation improves the system's schedulability. In this\u0000paper, we propose a hybrid multi-layer design space exploration technique to\u0000solve this multi-resource management problem. We explore the interplay between\u0000cache partitioning and schedulability by systematically interleaving three\u0000optimization layers, viz., (i) in the outer layer, we perform a breadth-first\u0000search combined with proactive pruning for cache partitioning; (ii) in the\u0000middle layer, we exploit a first-fit heuristic for allocating tasks to cores;\u0000and (iii) in the inner layer, we use the well-known recurrence relation for the\u0000schedulability analysis of non-preemptive fixed-priority (NP-FP) tasks in a\u0000uniprocessor setting. Although our focus is on NP-FP scheduling, we evaluate\u0000the flexibility of our framework in supporting different scheduling policies\u0000(NP-EDF, P-EDF) by plugging in appropriate analysis methods in the inner layer.\u0000Experiments show that, compared to the state-of-the-art techniques, the\u0000proposed framework can improve the real-time schedulability of NP-FP task sets\u0000by an average of 15.2% with a maximum improvement of 233.6% (when tasks are\u0000highly cache-sensitive) and a minimum of 1.6% (when cache sensitivity is low).\u0000For such task sets, we found that clustering similar-period (or mutually\u0000compatible) tasks often leads to higher schedulability (on average 7.6%) than\u0000clustering by cache sensitivity. In our evaluation, the framework also achieves\u0000good results for preemptive and dynamic-priority scheduling policies.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"17 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138542516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Persistent Memory File Systems: A Survey 持久性内存文件系统:综述
Pub Date : 2023-10-04 DOI: arxiv-2310.02880
Wiebe van Breukelen, Animesh Trivedi
Persistent Memory (PM) is non-volatile byte-addressable memory that offersread and write latencies in the order of magnitude smaller than flash storage,such as SSDs. This survey discusses how file systems address the most prominentchallenges in the implementation of file systems for Persistent Memory. First,we discuss how the properties of Persistent Memory change file system design.Second, we discuss work that aims to optimize small file I/O and the associatedmeta-data resolution. Third, we address how existing Persistent Memory filesystems achieve (meta) data persistence and consistency.
持久性内存(PM)是非易失的字节可寻址内存,它提供的读和写延迟比闪存(如ssd)小几个数量级。本调查讨论了文件系统如何解决持久性内存文件系统实现中最突出的挑战。首先,我们讨论持久性内存的属性如何改变文件系统的设计。其次,我们讨论了旨在优化小文件I/O和相关元数据分辨率的工作。第三,我们讨论了现有的Persistent Memory文件系统是如何实现(元)数据持久性和一致性的。
{"title":"Persistent Memory File Systems: A Survey","authors":"Wiebe van Breukelen, Animesh Trivedi","doi":"arxiv-2310.02880","DOIUrl":"https://doi.org/arxiv-2310.02880","url":null,"abstract":"Persistent Memory (PM) is non-volatile byte-addressable memory that offers\u0000read and write latencies in the order of magnitude smaller than flash storage,\u0000such as SSDs. This survey discusses how file systems address the most prominent\u0000challenges in the implementation of file systems for Persistent Memory. First,\u0000we discuss how the properties of Persistent Memory change file system design.\u0000Second, we discuss work that aims to optimize small file I/O and the associated\u0000meta-data resolution. Third, we address how existing Persistent Memory file\u0000systems achieve (meta) data persistence and consistency.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"53 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138542517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Case Study: Securing Embedded Linux Using CHERI 案例研究:使用CHERI保护嵌入式Linux
Pub Date : 2023-10-02 DOI: arxiv-2310.00933
Hesham Almatary
The current embedded Linux variant lacks security as it does not have or useMMU support. It does not also use MPUs as they do not fit with its softwaremodel because of the design drawbacks of MPUs (i.e., coarse-grained protectionwith fixed number of protected regions). We secure the existing embedded Linuxversion of the RISC-V port using CHERI. CHERI is hardware-softwarecapability-based system that leverages the ISA, toolchain, programminglanaguages, operating systems, and applications in order to provide completepointer and memory safety. We believe that CHERI could provide significantsecurity guarantees for high-end dynamic embedded systems at lower costs,compared to MMUs and MPUs, by: 1) building the entire software stack inpure-capability CHERI C mode which provides complete spatial memory safety atthe kernel and user-level, 2) isolating user programs as separate ELFs, eachwith its own CHERI-based capability table; this provides spatial memory safetysimilar to what the MMU offers (i.e., user programs cannot access each other'smemory), 3) isolating user programs from the kernel as the kernel has its owncapability table from the users and vice versa, and 4) compartmentalisingkernel modules using CompartOS' linkage-based compartmentalisation. This offersa new security front that is not possible using the current MMU-based Linux,where vulnerable/malicious kernel modules (e.g., device drivers) executing inthe kernel space would not compromise or take down the entire system. These arethe four main contributions of this paper, presenting novel CHERI-basedmechanisms to secure embedded Linux.
当前的嵌入式Linux变体缺乏安全性,因为它不支持或不使用emmu支持。它也不使用微处理器,因为它们不适合它的软件模型,因为微处理器的设计缺陷(即,具有固定数量的受保护区域的粗粒度保护)。我们使用CHERI保护现有的嵌入式linux版本的RISC-V端口。CHERI是一种基于硬件和软件可重用性的系统,它利用ISA、工具链、编程语言、操作系统和应用程序来提供完整的指针和内存安全性。我们相信,与mmu和mpu相比,CHERI可以以更低的成本为高端动态嵌入式系统提供重要的安全保障,通过:1)构建整个软件堆栈的全功能CHERI C模式,在内核和用户级提供完整的空间存储安全;2)将用户程序隔离为独立的elf,每个elf都有自己的基于CHERI的能力表;这提供了类似于MMU提供的空间内存安全性(即,用户程序不能访问彼此的内存),3)将用户程序从内核中隔离出来,因为内核从用户中有自己的能力表,反之亦然,4)使用CompartOS的基于链接的划分将内核模块划分。这提供了使用当前基于mmu的Linux无法实现的新的安全前线,在内核空间中执行的易受攻击/恶意内核模块(例如,设备驱动程序)不会危及或摧毁整个系统。这是本文的四个主要贡献,提出了新的基于cheri的机制来保护嵌入式Linux。
{"title":"Case Study: Securing Embedded Linux Using CHERI","authors":"Hesham Almatary","doi":"arxiv-2310.00933","DOIUrl":"https://doi.org/arxiv-2310.00933","url":null,"abstract":"The current embedded Linux variant lacks security as it does not have or use\u0000MMU support. It does not also use MPUs as they do not fit with its software\u0000model because of the design drawbacks of MPUs (i.e., coarse-grained protection\u0000with fixed number of protected regions). We secure the existing embedded Linux\u0000version of the RISC-V port using CHERI. CHERI is hardware-software\u0000capability-based system that leverages the ISA, toolchain, programming\u0000lanaguages, operating systems, and applications in order to provide complete\u0000pointer and memory safety. We believe that CHERI could provide significant\u0000security guarantees for high-end dynamic embedded systems at lower costs,\u0000compared to MMUs and MPUs, by: 1) building the entire software stack in\u0000pure-capability CHERI C mode which provides complete spatial memory safety at\u0000the kernel and user-level, 2) isolating user programs as separate ELFs, each\u0000with its own CHERI-based capability table; this provides spatial memory safety\u0000similar to what the MMU offers (i.e., user programs cannot access each other's\u0000memory), 3) isolating user programs from the kernel as the kernel has its own\u0000capability table from the users and vice versa, and 4) compartmentalising\u0000kernel modules using CompartOS' linkage-based compartmentalisation. This offers\u0000a new security front that is not possible using the current MMU-based Linux,\u0000where vulnerable/malicious kernel modules (e.g., device drivers) executing in\u0000the kernel space would not compromise or take down the entire system. These are\u0000the four main contributions of this paper, presenting novel CHERI-based\u0000mechanisms to secure embedded Linux.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"25 2","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138522168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The First Principles of Big Memory Systems 大内存系统的第一原理
Pub Date : 2023-09-30 DOI: arxiv-2310.00428
Yu Hua
In this paper, we comprehensively analyze the vertical and horizontalextensions of existing memory hierarchy. The difference between memory and bigmemory is well reported. We present the state-of-the-art studies upon the bigmemory systems, together with design methodology and implementations.Persistence is the first principle of big memory systems. We further show thefull-stack and moving persistence.
在本文中,我们全面分析了现有内存层次结构的垂直和水平扩展。内存和大内存之间的区别有很好的报道。我们介绍了大内存系统的最新研究,以及设计方法和实现。持久性是大内存系统的首要原则。我们进一步展示了全栈和移动持久性。
{"title":"The First Principles of Big Memory Systems","authors":"Yu Hua","doi":"arxiv-2310.00428","DOIUrl":"https://doi.org/arxiv-2310.00428","url":null,"abstract":"In this paper, we comprehensively analyze the vertical and horizontal\u0000extensions of existing memory hierarchy. The difference between memory and big\u0000memory is well reported. We present the state-of-the-art studies upon the big\u0000memory systems, together with design methodology and implementations.\u0000Persistence is the first principle of big memory systems. We further show the\u0000full-stack and moving persistence.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"40 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138522090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Operating Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1