首页 > 最新文献

2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)最新文献

英文 中文
Bandwidth Scheduling with Flexible Multi-paths in High-Performance Networks 高性能网络中灵活多路径的带宽调度
Xiaoyang Zhang, C. Wu, Liudong Zuo, Aiqin Hou, Yongqiang Wang
Modern data-intensive applications require the transfer of big data over high-performance networks (HPNs) through bandwidth reservation for various purposes such as data storage and analysis. The key performance metrics for bandwidth scheduling include the utilization of network resources and the satisfaction of user requests. In this paper, for a given batch of Deadline-Constrained Bandwidth Reservation Requests (DCBRRs), we attempt to maximize the number of satisfied requests with flexible scheduling options over link-disjoint paths in an HPN while achieving the best average Earliest Completion Time (ECT) or Shortest Duration (SD) of scheduled requests. We further consider this problem from two bandwidth-oriented principles: (i) Minimum Bandwidth Principle (MINBP), and (ii) Maximum Bandwidth Principle (MAXBP). We show that both of these problem variants are NP-complete, and propose two heuristic algorithms with polynomial-time complexity for each. We conduct bandwidth scheduling experiments on both small-and large-scale DCBRRs in a real-life HPN topology for performance comparison. Extensive results show the superiority of the proposed algorithms over existing ones in comparison.
现代数据密集型应用需要通过预留带宽在高性能网络(HPNs)上传输大数据,以实现数据存储和分析等各种目的。带宽调度的关键性能指标包括网络资源的利用率和用户请求的满意度。在本文中,对于给定的一批受截止日期约束的带宽保留请求(dbrr),我们尝试在HPN中链路不连接路径上使用灵活的调度选项来最大化满足请求的数量,同时获得调度请求的最佳平均最早完成时间(ECT)或最短持续时间(SD)。我们从两个面向带宽的原则:(i)最小带宽原则(MINBP)和(ii)最大带宽原则(MAXBP)进一步考虑这个问题。我们证明了这两个问题变体都是np完全的,并为每个变体提出了两个具有多项式时间复杂度的启发式算法。我们在真实的HPN拓扑中对小型和大型dcrr进行了带宽调度实验,以进行性能比较。大量的实验结果表明,本文提出的算法与现有算法相比具有优越性。
{"title":"Bandwidth Scheduling with Flexible Multi-paths in High-Performance Networks","authors":"Xiaoyang Zhang, C. Wu, Liudong Zuo, Aiqin Hou, Yongqiang Wang","doi":"10.1109/CCGRID.2018.00-69","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00-69","url":null,"abstract":"Modern data-intensive applications require the transfer of big data over high-performance networks (HPNs) through bandwidth reservation for various purposes such as data storage and analysis. The key performance metrics for bandwidth scheduling include the utilization of network resources and the satisfaction of user requests. In this paper, for a given batch of Deadline-Constrained Bandwidth Reservation Requests (DCBRRs), we attempt to maximize the number of satisfied requests with flexible scheduling options over link-disjoint paths in an HPN while achieving the best average Earliest Completion Time (ECT) or Shortest Duration (SD) of scheduled requests. We further consider this problem from two bandwidth-oriented principles: (i) Minimum Bandwidth Principle (MINBP), and (ii) Maximum Bandwidth Principle (MAXBP). We show that both of these problem variants are NP-complete, and propose two heuristic algorithms with polynomial-time complexity for each. We conduct bandwidth scheduling experiments on both small-and large-scale DCBRRs in a real-life HPN topology for performance comparison. Extensive results show the superiority of the proposed algorithms over existing ones in comparison.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"92 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126204385","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Efficient Messaging for Java Applications Running in Data Centers 在数据中心中运行的Java应用程序的高效消息传递
Kevin Beineke, Stefan Nothaas, M. Schöttner
Big data and large-scale Java applications often aggregate the resources of many servers. Low-latency and high-throughput network communication is important, if the applications have to process many concurrent interactive queries. We designed DXNet to address these challenges providing fast object de-/serialization, automatic connection management and zero-copy messaging. The latter includes sending of asynchronous messages as well as synchronous requests/responses and an event-driven message receiving approach. DXNet is optimized for small messages (< 64 bytes) in order to support highly interactive web applications, e.g., graph-based information retrieval, but works well with larger messages (e.g., 8 MB) as well. DXNet is available as standalone component on Github and its modular design is open for different transports currently supporting Ethernet and InfiniBand. The evaluation with micro benchmarks and YCSB using Ethernet and InfiniBand shows request-response latencies sub 10 µs (round-trip) including object de-/serialization, as well as a maximum throughput of more than 9 GByte/s.
大数据和大规模Java应用程序通常会聚合许多服务器的资源。如果应用程序必须处理许多并发交互式查询,那么低延迟和高吞吐量的网络通信非常重要。我们设计DXNet来解决这些挑战,提供快速对象反序列化、自动连接管理和零拷贝消息传递。后者包括异步消息的发送以及同步请求/响应和事件驱动的消息接收方法。DXNet针对小消息(< 64字节)进行了优化,以便支持高度交互的web应用程序,例如,基于图的信息检索,但也可以很好地处理较大的消息(例如,8 MB)。DXNet作为独立组件在Github上可用,其模块化设计对目前支持以太网和InfiniBand的不同传输开放。使用以太网和InfiniBand的微基准测试和YCSB的评估显示,请求-响应延迟低于10µs(往返),包括对象反/序列化,以及超过9 GByte/s的最大吞吐量。
{"title":"Efficient Messaging for Java Applications Running in Data Centers","authors":"Kevin Beineke, Stefan Nothaas, M. Schöttner","doi":"10.1109/CCGRID.2018.00090","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00090","url":null,"abstract":"Big data and large-scale Java applications often aggregate the resources of many servers. Low-latency and high-throughput network communication is important, if the applications have to process many concurrent interactive queries. We designed DXNet to address these challenges providing fast object de-/serialization, automatic connection management and zero-copy messaging. The latter includes sending of asynchronous messages as well as synchronous requests/responses and an event-driven message receiving approach. DXNet is optimized for small messages (< 64 bytes) in order to support highly interactive web applications, e.g., graph-based information retrieval, but works well with larger messages (e.g., 8 MB) as well. DXNet is available as standalone component on Github and its modular design is open for different transports currently supporting Ethernet and InfiniBand. The evaluation with micro benchmarks and YCSB using Ethernet and InfiniBand shows request-response latencies sub 10 µs (round-trip) including object de-/serialization, as well as a maximum throughput of more than 9 GByte/s.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129483147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
A Low-Latency Memory-Efficient IPv6 Lookup Engine Implemented on FPGA Using High-Level Synthesis 一种基于FPGA的低延迟内存高效IPv6查找引擎
Thibaut Stimpfling, J. Langlois, N. Bélanger, Y. Savaria
The emergence of 5G networks and real-time applications across networks has a strong impact on the performance requirements of IP lookup engines. These engines must support not only high-bandwidth but also low-latency lookup operations. This paper presents the hardware architecture of a low-latency IPv6 lookup engine capable of supporting the bandwidth of current Ethernet links. The engine implements the SHIP lookup algorithm, which exploits prefix characteristics to build a compact and scalable data structure. The proposed hardware architecture leverages the characteristics of the data structure to support low-latency lookup operations, while making efficient use of memory. The architecture is described in C++, synthesized with a high-level synthesis tool, then implemented on a Virtex-7 FPGA. Compared to other well-known approaches, the proposed IPvThe emergence of 5G networks and real-time applications across networks has a strong impact on the performance requirements of IP lookup engines. These engines must support not only high-bandwidth but also low-latency lookup operations. This paper presents the hardware architecture of a low-latency IPv6 lookup engine capable of supporting the bandwidth of current Ethernet links. The engine implements the SHIP lookup algorithm, which exploits prefix characteristics to build a compact and scalable data structure. The proposed hardware architecture leverages the characteristics of the data structure to support lowlatency lookup operations, while making efficient use of memory. The architecture is described in C++, synthesized with a highlevel synthesis tool, then implemented on a Virtex-7 FPGA. Compared to the proposed IPv6 lookup architecture, other wellknown approaches use at least 87% more memory per prefix, while increasing the lookup latency by a factor of 2.3×.6 lookup architecture reduces lookup latency by a factor of 2.3x and uses as much as 46% less memory per prefix for a synthetic prefix table holding 580 k entries.
5G网络和跨网实时应用的出现,对IP查找引擎的性能要求产生了强烈的影响。这些引擎不仅要支持高带宽,还要支持低延迟查找操作。本文提出了一种能够支持当前以太网链路带宽的低延迟IPv6查找引擎的硬件架构。该引擎实现SHIP查找算法,该算法利用前缀特征构建紧凑且可扩展的数据结构。所建议的硬件体系结构利用数据结构的特性来支持低延迟查找操作,同时有效地利用内存。用c++语言描述了该体系结构,利用高级综合工具进行了综合,然后在Virtex-7 FPGA上实现了该体系结构。与其他众所周知的方法相比,5G网络和跨网络实时应用的出现对IP查找引擎的性能要求产生了强烈的影响。这些引擎不仅要支持高带宽,还要支持低延迟查找操作。本文提出了一种能够支持当前以太网链路带宽的低延迟IPv6查找引擎的硬件架构。该引擎实现SHIP查找算法,该算法利用前缀特征构建紧凑且可扩展的数据结构。所建议的硬件体系结构利用数据结构的特性来支持低延迟查找操作,同时有效地利用内存。用c++语言描述了该体系结构,并利用高级综合工具进行了综合,最后在Virtex-7 FPGA上实现。与提出的IPv6查找架构相比,其他已知的方法在每个前缀上至少多使用87%的内存,同时将查找延迟增加2.3倍。6查找架构将查找延迟减少了2.3倍,并且对于包含580 k项的合成前缀表,每个前缀使用的内存减少了46%。
{"title":"A Low-Latency Memory-Efficient IPv6 Lookup Engine Implemented on FPGA Using High-Level Synthesis","authors":"Thibaut Stimpfling, J. Langlois, N. Bélanger, Y. Savaria","doi":"10.1109/CCGRID.2018.00067","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00067","url":null,"abstract":"The emergence of 5G networks and real-time applications across networks has a strong impact on the performance requirements of IP lookup engines. These engines must support not only high-bandwidth but also low-latency lookup operations. This paper presents the hardware architecture of a low-latency IPv6 lookup engine capable of supporting the bandwidth of current Ethernet links. The engine implements the SHIP lookup algorithm, which exploits prefix characteristics to build a compact and scalable data structure. The proposed hardware architecture leverages the characteristics of the data structure to support low-latency lookup operations, while making efficient use of memory. The architecture is described in C++, synthesized with a high-level synthesis tool, then implemented on a Virtex-7 FPGA. Compared to other well-known approaches, the proposed IPvThe emergence of 5G networks and real-time applications across networks has a strong impact on the performance requirements of IP lookup engines. These engines must support not only high-bandwidth but also low-latency lookup operations. This paper presents the hardware architecture of a low-latency IPv6 lookup engine capable of supporting the bandwidth of current Ethernet links. The engine implements the SHIP lookup algorithm, which exploits prefix characteristics to build a compact and scalable data structure. The proposed hardware architecture leverages the characteristics of the data structure to support lowlatency lookup operations, while making efficient use of memory. The architecture is described in C++, synthesized with a highlevel synthesis tool, then implemented on a Virtex-7 FPGA. Compared to the proposed IPv6 lookup architecture, other wellknown approaches use at least 87% more memory per prefix, while increasing the lookup latency by a factor of 2.3×.6 lookup architecture reduces lookup latency by a factor of 2.3x and uses as much as 46% less memory per prefix for a synthetic prefix table holding 580 k entries.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134535923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CloudRanger: Root Cause Identification for Cloud Native Systems CloudRanger:云原生系统的根本原因识别
Ping Wang, Jingmin Xu, Meng Ma, Weilan Lin, Disheng Pan, Y. Wang, Pengfei Chen
As more and more systems are migrating to cloud environment, the cloud native system becomes a trend. This paper presents the challenges and implications when diagnosing root causes for cloud native systems by analyzing some real incidents occurred in IBM Bluemix (a large commercial cloud). To tackle these challenges, we propose CloudRanger, a novel system dedicated for cloud native systems. To make our system more general, we propose a dynamic causal relationship analysis approach to construct impact graphs amongst applications without given the topology. A heuristic investigation algorithm based on second-order random walk is proposed to identify the culprit services which are responsible for cloud incidents. Experimental results in both simulation environment and IBM Bluemix platform show that CloudRanger outperforms some state-of-the-art approaches with a 10% improvement in accuracy. It offers a fast identification of culprit services when an anomaly occurs. Moreover, this system can be deployed rapidly and easily in multiple kinds of cloud native systems without any predefined knowledge.
随着越来越多的系统向云环境迁移,云原生系统成为一种趋势。本文通过分析IBM Bluemix(一个大型商业云)中发生的一些真实事件,介绍了在诊断云原生系统的根本原因时所面临的挑战和影响。为了应对这些挑战,我们提出了CloudRanger,一个专门用于云原生系统的新系统。为了使我们的系统更具通用性,我们提出了一种动态因果关系分析方法,在不给定拓扑的情况下构建应用程序之间的影响图。提出了一种基于二阶随机漫步的启发式调查算法,用于识别导致云事件的罪魁祸首服务。在模拟环境和IBM Bluemix平台上的实验结果表明,CloudRanger的准确率比一些最先进的方法提高了10%。当异常发生时,它提供了对罪魁祸首服务的快速识别。此外,该系统可以快速、轻松地部署在多种云原生系统中,无需任何预先定义的知识。
{"title":"CloudRanger: Root Cause Identification for Cloud Native Systems","authors":"Ping Wang, Jingmin Xu, Meng Ma, Weilan Lin, Disheng Pan, Y. Wang, Pengfei Chen","doi":"10.1109/CCGRID.2018.00076","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00076","url":null,"abstract":"As more and more systems are migrating to cloud environment, the cloud native system becomes a trend. This paper presents the challenges and implications when diagnosing root causes for cloud native systems by analyzing some real incidents occurred in IBM Bluemix (a large commercial cloud). To tackle these challenges, we propose CloudRanger, a novel system dedicated for cloud native systems. To make our system more general, we propose a dynamic causal relationship analysis approach to construct impact graphs amongst applications without given the topology. A heuristic investigation algorithm based on second-order random walk is proposed to identify the culprit services which are responsible for cloud incidents. Experimental results in both simulation environment and IBM Bluemix platform show that CloudRanger outperforms some state-of-the-art approaches with a 10% improvement in accuracy. It offers a fast identification of culprit services when an anomaly occurs. Moreover, this system can be deployed rapidly and easily in multiple kinds of cloud native systems without any predefined knowledge.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130277859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 71
SHMEMGraph: Efficient and Balanced Graph Processing Using One-Sided Communication SHMEMGraph:使用单侧通信的高效和平衡的图处理
Huansong Fu, Manjunath Gorentla Venkata, Shaeke Salman, N. Imam, Weikuan Yu
State-of-the-art synchronous graph processing frameworks face both inefficiency and imbalance issues that cause their performance to be suboptimal. These issues include the inefficiency of communication and the imbalanced graph computation/communication costs in an iteration. We propose to replace their conventional two-sided communication model with the one-sided counterpart. Accordingly, we design SHMEMGraph, an efficient and balanced graph processing framework that is formulated across a global memory space and takes advantage of the flexibility and efficiency of one-sided communication for graph processing. Through an efficient one-sided communication channel, SHMEMGraph utilizes the high-performance operations with RDMA while minimizing the resource contention within a computer node. In addition, SHMEMGraph synthesizes a number of optimizations to address both computation imbalance and communication imbalance. By using a graph of 1 billion edges, our evaluation shows that compared to the state-of-the-art Gemini framework, SHMEMGraph achieves an average improvement of 35.5% in terms of job completion time for five representative graph algorithms.
最先进的同步图形处理框架面临着低效和不平衡的问题,导致它们的性能不是最优的。这些问题包括通信效率低下和迭代中不平衡的图计算/通信成本。我们建议用单边通信模式取代传统的双边通信模式。因此,我们设计了SHMEMGraph,这是一个高效和平衡的图形处理框架,它在全局内存空间中制定,并利用单侧通信的灵活性和效率进行图形处理。通过高效的单侧通信通道,SHMEMGraph利用了RDMA的高性能操作,同时最大限度地减少了计算机节点内的资源争用。此外,SHMEMGraph综合了许多优化来解决计算不平衡和通信不平衡。通过使用10亿个边的图,我们的评估表明,与最先进的Gemini框架相比,SHMEMGraph在五个代表性图算法的任务完成时间方面平均提高了35.5%。
{"title":"SHMEMGraph: Efficient and Balanced Graph Processing Using One-Sided Communication","authors":"Huansong Fu, Manjunath Gorentla Venkata, Shaeke Salman, N. Imam, Weikuan Yu","doi":"10.1109/CCGRID.2018.00078","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00078","url":null,"abstract":"State-of-the-art synchronous graph processing frameworks face both inefficiency and imbalance issues that cause their performance to be suboptimal. These issues include the inefficiency of communication and the imbalanced graph computation/communication costs in an iteration. We propose to replace their conventional two-sided communication model with the one-sided counterpart. Accordingly, we design SHMEMGraph, an efficient and balanced graph processing framework that is formulated across a global memory space and takes advantage of the flexibility and efficiency of one-sided communication for graph processing. Through an efficient one-sided communication channel, SHMEMGraph utilizes the high-performance operations with RDMA while minimizing the resource contention within a computer node. In addition, SHMEMGraph synthesizes a number of optimizations to address both computation imbalance and communication imbalance. By using a graph of 1 billion edges, our evaluation shows that compared to the state-of-the-art Gemini framework, SHMEMGraph achieves an average improvement of 35.5% in terms of job completion time for five representative graph algorithms.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132445599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
AKIN: A Streaming Graph Partitioning Algorithm for Distributed Graph Storage Systems 分布式图存储系统的流图分区算法
Wei Zhang, Yong Chen, Dong Dai
Many graph-related applications face the challenge of managing excessive and ever-growing graph data in a distributed environment. Therefore, it is necessary to consider a graph partitioning algorithm to distribute graph data onto multiple machines as the data comes in. Balancing data distribution and minimizing edge-cut ratio are two basic pursuits of the graph partitioning problem. While achieving balanced partitions for streaming graphs is easy, existing graph partitioning algorithms either fail to work on streaming workloads, or leave edge-cut ratio to be further improved. Our research aims to provide a better solution that fits the need of streaming graph partitioning in a distributed system, which further reduces the edge-cut ratio while maintaining rough balance among all partitions. We exploit the similarity measure on the degree of vertices to gather structuralrelated vertices in the same partition as much as possible, this reduces the edge-cut ratio even further as compared to the state-of-the-art streaming graph partitioning algorithm - FENNEL. Our evaluation shows that our streaming graph partitioning algorithm is able to achieve better partitioning quality in terms of edge-cut ratio (up to 20% reduction as compared to FENNEL) while maintaining decent balance between all partitions, and such improvement applies to various real-life graphs.
许多与图形相关的应用程序都面临着在分布式环境中管理过多且不断增长的图形数据的挑战。因此,有必要考虑一种图分区算法,以便在数据传入时将图数据分布到多台机器上。平衡数据分布和最小化切边比是图划分问题的两个基本追求。虽然实现流图的均衡分区很容易,但现有的图分区算法要么无法处理流工作负载,要么将切边率留作进一步改进。我们的研究旨在提供一种更好的解决方案,以适应分布式系统中流图分区的需要,在保持所有分区之间大致平衡的同时进一步降低切边率。我们利用顶点度的相似性度量来尽可能多地收集同一分区中与结构相关的顶点,与最先进的流图分区算法FENNEL相比,这进一步降低了边缘切割率。我们的评估表明,我们的流图分区算法能够在切边率方面实现更好的分区质量(与FENNEL相比减少了20%),同时保持所有分区之间的良好平衡,这种改进适用于各种现实生活中的图。
{"title":"AKIN: A Streaming Graph Partitioning Algorithm for Distributed Graph Storage Systems","authors":"Wei Zhang, Yong Chen, Dong Dai","doi":"10.1109/CCGRID.2018.00033","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00033","url":null,"abstract":"Many graph-related applications face the challenge of managing excessive and ever-growing graph data in a distributed environment. Therefore, it is necessary to consider a graph partitioning algorithm to distribute graph data onto multiple machines as the data comes in. Balancing data distribution and minimizing edge-cut ratio are two basic pursuits of the graph partitioning problem. While achieving balanced partitions for streaming graphs is easy, existing graph partitioning algorithms either fail to work on streaming workloads, or leave edge-cut ratio to be further improved. Our research aims to provide a better solution that fits the need of streaming graph partitioning in a distributed system, which further reduces the edge-cut ratio while maintaining rough balance among all partitions. We exploit the similarity measure on the degree of vertices to gather structuralrelated vertices in the same partition as much as possible, this reduces the edge-cut ratio even further as compared to the state-of-the-art streaming graph partitioning algorithm - FENNEL. Our evaluation shows that our streaming graph partitioning algorithm is able to achieve better partitioning quality in terms of edge-cut ratio (up to 20% reduction as compared to FENNEL) while maintaining decent balance between all partitions, and such improvement applies to various real-life graphs.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122249496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Improving Data Integrity in Linux Software RAID with Protection Information (T10-PI) 利用保护信息提高Linux软件RAID的数据完整性(T10-PI)
Baoquan Zhang, R. Rajachandrasekar, Alireza Haghdoost, Lance Evans, D. Du
The T10 DIF (Data Integrity Field) and DIX (Data Integrity Extension) specifications provide mechanisms to guarantee end-to-end data integrity and protection in the face of silent data corruption in modern storage systems. However, the Multiple Devices (MD) software RAID driver in Linux does not fully leverage these capabilities to provide such end-to-end guarantees with widely-used RAID modes such as 5 and 6, thereby causing an "integrity gap" in the Linux I/O stack. This paper describes the design and performance characteristics of a DIX-aware MD module that plugs this integrity gap with minimal overhead to client applications. A PI (Protection Information) operator is added in MD to handle the PI-related operations, and dedicated buffers for PI are allocated and managed in MD RAID-5/6 personality's stripe structures to generate, store, and verify the PI. This allows seamless exchange of PI information among end-applications running in user mode, file systems, the linux block layer, and PI-capable HBAs and drives. Our evaluations show that the DIX-aware MD module has the capability of detecting SDC with the tolerable performance penalty.
T10 DIF(数据完整性字段)和DIX(数据完整性扩展)规范提供了在现代存储系统中面对无声数据损坏时保证端到端数据完整性和保护的机制。然而,Linux中的multi Devices (MD)软件RAID驱动程序并没有充分利用这些功能,为广泛使用的RAID模式(如5和6)提供端到端保证,从而导致Linux I/O堆栈中的“完整性缺口”。本文描述了一个dix感知MD模块的设计和性能特征,该模块以最小的开销为客户端应用程序填补了这种完整性缺口。在MD中增加一个PI(保护信息)操作符来处理PI相关操作,并在MD RAID-5/6人格的条带结构中为PI分配和管理专用缓冲区,以生成、存储和验证PI。这允许在以用户模式运行的终端应用程序、文件系统、linux块层和支持PI的hba和驱动器之间无缝交换PI信息。我们的评估表明,dix感知MD模块具有在可容忍的性能损失下检测SDC的能力。
{"title":"Improving Data Integrity in Linux Software RAID with Protection Information (T10-PI)","authors":"Baoquan Zhang, R. Rajachandrasekar, Alireza Haghdoost, Lance Evans, D. Du","doi":"10.1109/CCGRID.2018.00091","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00091","url":null,"abstract":"The T10 DIF (Data Integrity Field) and DIX (Data Integrity Extension) specifications provide mechanisms to guarantee end-to-end data integrity and protection in the face of silent data corruption in modern storage systems. However, the Multiple Devices (MD) software RAID driver in Linux does not fully leverage these capabilities to provide such end-to-end guarantees with widely-used RAID modes such as 5 and 6, thereby causing an \"integrity gap\" in the Linux I/O stack. This paper describes the design and performance characteristics of a DIX-aware MD module that plugs this integrity gap with minimal overhead to client applications. A PI (Protection Information) operator is added in MD to handle the PI-related operations, and dedicated buffers for PI are allocated and managed in MD RAID-5/6 personality's stripe structures to generate, store, and verify the PI. This allows seamless exchange of PI information among end-applications running in user mode, file systems, the linux block layer, and PI-capable HBAs and drives. Our evaluations show that the DIX-aware MD module has the capability of detecting SDC with the tolerable performance penalty.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127116171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RISP: A Reconfigurable In-Storage Processing Framework with Energy-Awareness RISP:具有能量感知的可重构存储处理框架
Xiaojia Song, T. Xie, Wen Pan
Existing in-storage processing (ISP) techniques mainly focus on maximizing data processing rate by always utilizing total storage data processing resources for all applications. We find that this "always running in full gear" strategy wastes energy for some applications with a low data processing complexity. In this paper we propose RISP (Reconfigurable ISP), an energy-aware reconfigurable ISP framework that employs FPGA as data processing cells and NVM controllers. It can reconfigure storage data processing resources to achieve a high energy-efficiency without any performance degradation for big data analysis applications. RISP is modeled and then validated on an FPGA board. Experimental results show that compared with traditional host-CPU based computing RISP (with 16 channels or more) improves performance by 1.6-25.4× while saving energy by a factor of 2.2-161. Further, its reconfigurability can provide up to 77.2% additional energy saving by judiciously enabling data processing resources that are sufficient for an application.
现有的存储内处理(ISP)技术主要是通过对所有应用程序始终使用全部存储数据处理资源来实现数据处理速率的最大化。我们发现这种“总是全速运行”的策略对于一些数据处理复杂性较低的应用程序来说浪费了能源。在本文中,我们提出了一种能量感知的可重构ISP框架RISP(可重构ISP),它采用FPGA作为数据处理单元和NVM控制器。它可以重新配置存储数据处理资源,在不降低大数据分析应用性能的情况下实现高能效。建立了RISP模型,并在FPGA板上进行了验证。实验结果表明,与传统的基于主机- cpu的计算相比,RISP(16通道及以上)性能提高1.6-25.4倍,节能2.2-161倍。此外,通过明智地启用对应用程序足够的数据处理资源,其可重构性可以提供高达77.2%的额外节能。
{"title":"RISP: A Reconfigurable In-Storage Processing Framework with Energy-Awareness","authors":"Xiaojia Song, T. Xie, Wen Pan","doi":"10.1109/CCGRID.2018.00034","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00034","url":null,"abstract":"Existing in-storage processing (ISP) techniques mainly focus on maximizing data processing rate by always utilizing total storage data processing resources for all applications. We find that this \"always running in full gear\" strategy wastes energy for some applications with a low data processing complexity. In this paper we propose RISP (Reconfigurable ISP), an energy-aware reconfigurable ISP framework that employs FPGA as data processing cells and NVM controllers. It can reconfigure storage data processing resources to achieve a high energy-efficiency without any performance degradation for big data analysis applications. RISP is modeled and then validated on an FPGA board. Experimental results show that compared with traditional host-CPU based computing RISP (with 16 channels or more) improves performance by 1.6-25.4× while saving energy by a factor of 2.2-161. Further, its reconfigurability can provide up to 77.2% additional energy saving by judiciously enabling data processing resources that are sufficient for an application.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129931795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Location, Location, Location: Exploring Amazon EC2 Spot Instance Pricing Across Geographical Regions 位置、位置、位置:探索跨地理区域的Amazon EC2现货实例定价
Nnamdi Ekwe-Ekwe, A. Barker
Cloud computing is a ubiquitous part of the computing landscape. For many companies today, moving their computing infrastructure to the cloud reduces time to deployment and saves money. Spot Instances, a subset of Amazon's cloud computing infrastructure (EC2), expands upon this. They allow a user to bid on spare compute capacity in EC2 at heavily discounted prices. If other bids for the spare capacity exceeds the user's own, the user's instance is terminated. In this paper, we conduct one of the first detailed analyses of how location affects the overall cost of deployment of a Spot Instance. We analyse pricing data across all available AWS regions for 60 days for a variety of Spot Instances. We relate the pricing data we find to the overall AWS region and examine any patterns we see across the week. We find that location plays a critical role in Spot Instance pricing and that pricing differs, sometimes markedly, from region to region. We conclude by showing that it is indeed possible to run workloads on Spot Instances with low risk of termination and a low overall cost.
云计算是计算领域中无处不在的一部分。对于今天的许多公司来说,将他们的计算基础设施迁移到云端减少了部署时间并节省了资金。Spot Instances是Amazon云计算基础设施(EC2)的一个子集,在此基础上进行了扩展。它们允许用户以很大的折扣价格竞标EC2中的备用计算容量。如果其他对备用容量的出价超过用户自己的出价,则终止用户的实例。在本文中,我们对位置如何影响Spot实例部署的总体成本进行了首次详细分析。我们分析了所有可用AWS区域60天内各种现货实例的定价数据。我们将发现的定价数据与整个AWS区域联系起来,并检查我们在一周内看到的任何模式。我们发现,位置在现货实例定价中起着关键作用,并且定价有时会因地区而异,甚至显著不同。最后,我们展示了在Spot实例上运行工作负载具有低终止风险和低总体成本的可能性。
{"title":"Location, Location, Location: Exploring Amazon EC2 Spot Instance Pricing Across Geographical Regions","authors":"Nnamdi Ekwe-Ekwe, A. Barker","doi":"10.1109/CCGRID.2018.00059","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00059","url":null,"abstract":"Cloud computing is a ubiquitous part of the computing landscape. For many companies today, moving their computing infrastructure to the cloud reduces time to deployment and saves money. Spot Instances, a subset of Amazon's cloud computing infrastructure (EC2), expands upon this. They allow a user to bid on spare compute capacity in EC2 at heavily discounted prices. If other bids for the spare capacity exceeds the user's own, the user's instance is terminated. In this paper, we conduct one of the first detailed analyses of how location affects the overall cost of deployment of a Spot Instance. We analyse pricing data across all available AWS regions for 60 days for a variety of Spot Instances. We relate the pricing data we find to the overall AWS region and examine any patterns we see across the week. We find that location plays a critical role in Spot Instance pricing and that pricing differs, sometimes markedly, from region to region. We conclude by showing that it is indeed possible to run workloads on Spot Instances with low risk of termination and a low overall cost.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117066068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
TýrFS: Increasing Small Files Access Performance with Dynamic Metadata Replication TýrFS:通过动态元数据复制提高小文件访问性能
Pierre Matri, María S. Pérez, Alexandru Costan, Gabriel Antoniu
Small files are known to pose major performance challenges for file systems. Yet, such workloads are increasingly common in a number of Big Data Analytics workflows or large-scale HPC simulations. These challenges are mainly caused by the common architecture of most state-of-the-art file systems needing one or multiple metadata requests before being able to read from a file. Small input file size causes the overhead of this metadata management to gain relative importance as the size of each file decreases. In this paper we propose a set of techniques leveraging consistent hashing and dynamic metadata replication to significantly reduce this metadata overhead. We implement such techniques inside a new file system named TýrFS, built as a thin layer above the Týr object store. We prove that TýrFS increases small file access performance up to one order of magnitude compared to other state-of-the-art file systems, while only causing a minimal impact on file write throughput.
众所周知,小文件给文件系统带来了主要的性能挑战。然而,这种工作负载在许多大数据分析工作流程或大规模HPC模拟中越来越普遍。这些挑战主要是由于大多数最先进的文件系统的公共体系结构在能够从文件中读取之前需要一个或多个元数据请求。较小的输入文件大小导致元数据管理的开销随着每个文件大小的减小而变得相对重要。在本文中,我们提出了一组利用一致散列和动态元数据复制的技术,以显著减少元数据开销。我们在名为TýrFS的新文件系统中实现这些技术,该文件系统构建为Týr对象存储之上的瘦层。我们证明,与其他最先进的文件系统相比,TýrFS将小文件访问性能提高了一个数量级,同时只对文件写吞吐量造成最小的影响。
{"title":"TýrFS: Increasing Small Files Access Performance with Dynamic Metadata Replication","authors":"Pierre Matri, María S. Pérez, Alexandru Costan, Gabriel Antoniu","doi":"10.1109/CCGRID.2018.00072","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00072","url":null,"abstract":"Small files are known to pose major performance challenges for file systems. Yet, such workloads are increasingly common in a number of Big Data Analytics workflows or large-scale HPC simulations. These challenges are mainly caused by the common architecture of most state-of-the-art file systems needing one or multiple metadata requests before being able to read from a file. Small input file size causes the overhead of this metadata management to gain relative importance as the size of each file decreases. In this paper we propose a set of techniques leveraging consistent hashing and dynamic metadata replication to significantly reduce this metadata overhead. We implement such techniques inside a new file system named TýrFS, built as a thin layer above the Týr object store. We prove that TýrFS increases small file access performance up to one order of magnitude compared to other state-of-the-art file systems, while only causing a minimal impact on file write throughput.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124521707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1