首页 > 最新文献

2010 IEEE 16th International Conference on Parallel and Distributed Systems最新文献

英文 中文
Resilient Virtual Network Service Provision in Network Virtualization Environments 网络虚拟化环境下的弹性虚拟网络业务提供
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.26
Yang Chen, Jianxin Li, Tianyu Wo, Chunming Hu, Wantao Liu
Network Virtualization has recently emerged to provide scalable, customized and on-demand virtual network services over a shared substrate network. How to provide VN services with resiliency guarantees against network failures has become a critical issue, meanwhile the service resource usages should be minimized under the strict constraints such as link bandwidth capability and service resiliency guarantees etc. In this paper, we present a resource allocation algorithm to balance the tradeoff between service resource consumptions and service resiliency. By exploiting a heuristic VN mapping scheme and a restoration path selection scheme based on intelligent bandwidth sharing, the algorithm simultaneously makes cost-effective usage of network resources and protects VN services against network failures. We perform evaluations and find that the algorithm is near optimal in terms of network resource usage, especially the additional restoration bandwidth cost for resiliency protection.
网络虚拟化最近出现了,它在共享的基础网络上提供可伸缩的、定制的和按需的虚拟网络服务。如何在网络故障时为VN业务提供弹性保障已成为一个关键问题,同时在链路带宽能力和业务弹性保障等严格约束下,最大限度地减少业务资源的使用。在本文中,我们提出了一种资源分配算法来平衡服务资源消耗和服务弹性之间的权衡。该算法利用启发式VN映射方案和基于智能带宽共享的恢复路径选择方案,在有效利用网络资源的同时,保护VN业务免受网络故障的影响。我们进行了评估,发现该算法在网络资源使用方面接近最优,特别是弹性保护的额外恢复带宽成本。
{"title":"Resilient Virtual Network Service Provision in Network Virtualization Environments","authors":"Yang Chen, Jianxin Li, Tianyu Wo, Chunming Hu, Wantao Liu","doi":"10.1109/ICPADS.2010.26","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.26","url":null,"abstract":"Network Virtualization has recently emerged to provide scalable, customized and on-demand virtual network services over a shared substrate network. How to provide VN services with resiliency guarantees against network failures has become a critical issue, meanwhile the service resource usages should be minimized under the strict constraints such as link bandwidth capability and service resiliency guarantees etc. In this paper, we present a resource allocation algorithm to balance the tradeoff between service resource consumptions and service resiliency. By exploiting a heuristic VN mapping scheme and a restoration path selection scheme based on intelligent bandwidth sharing, the algorithm simultaneously makes cost-effective usage of network resources and protects VN services against network failures. We perform evaluations and find that the algorithm is near optimal in terms of network resource usage, especially the additional restoration bandwidth cost for resiliency protection.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"79 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130595969","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 54
Decentralized Search in Scale-Free P2P Networks 无标度P2P网络中的分散搜索
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.73
Praphul Chandra, D. Arora
Search in peer-to-peer networks is a challenging problem due to the absence of any centralized control & the limited information available at each node. When information is available about the overall structure of the network, use of this information can significantly improve the efficiency of decentralized search algorithms. Many peer-to-peer networks have been shown to exhibit power-law degree distribution. We propose two new decentralized search algorithms that can be used for efficient search in networks exhibiting scale-free design. Unlike previous work, our algorithms perform efficient search for a large range of power-law coefficients. Our algorithms are also unique in that they complete decentralized searches efficiently even when the network has disconnected components. As a corollary of this, our algorithms are also more resilient to network failure.
点对点网络中的搜索是一个具有挑战性的问题,因为缺乏任何集中控制和每个节点上可用的信息有限。当有关网络整体结构的信息可用时,使用这些信息可以显著提高分散搜索算法的效率。许多点对点网络已经显示出幂律度分布。我们提出了两种新的分散搜索算法,可用于在无标度设计的网络中进行有效搜索。与以前的工作不同,我们的算法对大范围的幂律系数进行有效的搜索。我们的算法也是独一无二的,即使在网络有断开连接的组件时,它们也能有效地完成分散的搜索。由此推论,我们的算法对网络故障也更有弹性。
{"title":"Decentralized Search in Scale-Free P2P Networks","authors":"Praphul Chandra, D. Arora","doi":"10.1109/ICPADS.2010.73","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.73","url":null,"abstract":"Search in peer-to-peer networks is a challenging problem due to the absence of any centralized control & the limited information available at each node. When information is available about the overall structure of the network, use of this information can significantly improve the efficiency of decentralized search algorithms. Many peer-to-peer networks have been shown to exhibit power-law degree distribution. We propose two new decentralized search algorithms that can be used for efficient search in networks exhibiting scale-free design. Unlike previous work, our algorithms perform efficient search for a large range of power-law coefficients. Our algorithms are also unique in that they complete decentralized searches efficiently even when the network has disconnected components. As a corollary of this, our algorithms are also more resilient to network failure.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130670725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
GMH: A Message Passing Toolkit for GPU Clusters GMH:一个GPU集群的消息传递工具包
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.35
Jie Chen, W. Watson, W. Mao
Driven by the market demand for high-definition 3D graphics, commodity graphics processing units (GPUs) have evolved into highly parallel, multi-threaded, many-core processors, which are ideal for data parallel computing. Many applications have been ported to run on a single GPU with tremendous speedups using general C-style programming languages such as CUDA. However, large applications require multiple GPUs and demand explicit message passing. This paper presents a message passing toolkit, called GMH (GPU Message Handler), on NVIDIA GPUs. This toolkit utilizes a data-parallel thread group as a way to map multiple GPUs on a single host to an MPI rank, and introduces a notion of virtual GPUs as a way to bind a thread to a GPU automatically. This toolkit provides high performance MPI style point-to-point and collective communication, but more importantly, facilitates event-driven APIs to allow an application to be managed and executed by the toolkit at runtime.
在市场对高清3D图形需求的推动下,商品图形处理单元(gpu)已经发展成为高度并行、多线程、多核处理器,这是数据并行计算的理想选择。许多应用程序已经被移植到单个GPU上运行,使用通用的c风格编程语言(如CUDA)具有巨大的速度。然而,大型应用程序需要多个gpu,并且需要显式的消息传递。本文提出了一种基于NVIDIA GPU的消息传递工具箱GMH (GPU message Handler)。该工具包利用数据并行线程组作为将单个主机上的多个GPU映射到MPI等级的方法,并引入虚拟GPU的概念,作为将线程自动绑定到GPU的方法。该工具包提供高性能MPI风格的点对点和集体通信,但更重要的是,它促进了事件驱动的api,从而允许工具包在运行时管理和执行应用程序。
{"title":"GMH: A Message Passing Toolkit for GPU Clusters","authors":"Jie Chen, W. Watson, W. Mao","doi":"10.1109/ICPADS.2010.35","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.35","url":null,"abstract":"Driven by the market demand for high-definition 3D graphics, commodity graphics processing units (GPUs) have evolved into highly parallel, multi-threaded, many-core processors, which are ideal for data parallel computing. Many applications have been ported to run on a single GPU with tremendous speedups using general C-style programming languages such as CUDA. However, large applications require multiple GPUs and demand explicit message passing. This paper presents a message passing toolkit, called GMH (GPU Message Handler), on NVIDIA GPUs. This toolkit utilizes a data-parallel thread group as a way to map multiple GPUs on a single host to an MPI rank, and introduces a notion of virtual GPUs as a way to bind a thread to a GPU automatically. This toolkit provides high performance MPI style point-to-point and collective communication, but more importantly, facilitates event-driven APIs to allow an application to be managed and executed by the toolkit at runtime.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123401831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
PinComm: Characterizing Intra-application Communication for the Many-Core Era PinComm:表征多核时代的应用内部通信
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.56
W. Heirman, D. Stroobandt, Narasinga Rao Miniskar, Roel Wuyts, F. Catthoor
As the number of cores in both embedded Multi-Processor Systems-on-Chip and general purpose processors keeps rising, on-chip communication becomes more and more important. In order to write efficient programs for these architectures it is therefore necessary to have a good idea of the communication behavior of an application. We present a communication profiler that extracts this behavior from compiled, sequential or parallel C/C++ programs, and constructs a dynamic data-flow graph at the level of major functional blocks. In contrast to existing methods of measuring inter-program communication, our tool automatically generates the program's data-flow graph and is less demanding for the developer. It can also be used to view differences between program phases (such as different video frames), which allows both input- and phase-specific optimizations to be made. We will also describe briefly how this information can subsequently be used to guide the effort of parallelizing the application, to co-design the software, memory hierarchy and communication hardware, and to provide new sources of communication-related runtime optimizations.
随着嵌入式多处理器片上系统和通用处理器内核数量的不断增加,片上通信变得越来越重要。因此,为了为这些体系结构编写高效的程序,有必要对应用程序的通信行为有一个很好的了解。我们提出了一个通信分析器,可以从编译的、顺序的或并行的C/ c++程序中提取这种行为,并在主要功能块级别构建一个动态数据流图。与现有的测量程序间通信的方法相比,我们的工具自动生成程序的数据流图,并且对开发人员的要求更低。它还可以用于查看程序阶段之间的差异(例如不同的视频帧),这允许进行特定于输入和阶段的优化。我们还将简要描述如何随后使用这些信息来指导并行化应用程序的工作,共同设计软件、内存层次结构和通信硬件,以及提供与通信相关的运行时优化的新来源。
{"title":"PinComm: Characterizing Intra-application Communication for the Many-Core Era","authors":"W. Heirman, D. Stroobandt, Narasinga Rao Miniskar, Roel Wuyts, F. Catthoor","doi":"10.1109/ICPADS.2010.56","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.56","url":null,"abstract":"As the number of cores in both embedded Multi-Processor Systems-on-Chip and general purpose processors keeps rising, on-chip communication becomes more and more important. In order to write efficient programs for these architectures it is therefore necessary to have a good idea of the communication behavior of an application. We present a communication profiler that extracts this behavior from compiled, sequential or parallel C/C++ programs, and constructs a dynamic data-flow graph at the level of major functional blocks. In contrast to existing methods of measuring inter-program communication, our tool automatically generates the program's data-flow graph and is less demanding for the developer. It can also be used to view differences between program phases (such as different video frames), which allows both input- and phase-specific optimizations to be made. We will also describe briefly how this information can subsequently be used to guide the effort of parallelizing the application, to co-design the software, memory hierarchy and communication hardware, and to provide new sources of communication-related runtime optimizations.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"131 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124202999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Towards a Common Interface for Overlay Network Simulators 面向覆盖网络模拟器的通用接口
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.33
C. Gross, Max Lehn, D. Stingl, A. Kovacevic, A. Buchmann, R. Steinmetz
Simulation has become an important evaluation method in the area of Peer-to-Peer (P2P) research due to the scalability limitations of evaluation test beds such as Planet Lab or G-Lab. Current simulators provide various abstraction levels for different underlay models, such that applications can be evaluated at different granularity. However, existing simulators suffer from a lack of interoperability and portability making the comparison of research results extremely difficult. To overcome this problem, we present an approach for a generic application interface for discrete-event P2P overlay network simulators. It enables porting of the same implementation of a targeted application once and then running it on various simulators as well as in a real network environment, thereby enabling a diverse and extensive evaluation. We established the feasibility of our approach and showed negligible memory and runtime overhead.
由于Planet Lab或G-Lab等评估试验台的可扩展性限制,仿真已成为P2P研究领域重要的评估方法。当前的模拟器为不同的底层模型提供了不同的抽象级别,这样应用程序就可以在不同的粒度上进行评估。然而,现有的模拟器缺乏互操作性和可移植性,使得研究结果的比较非常困难。为了克服这个问题,我们提出了一种离散事件P2P覆盖网络模拟器的通用应用接口方法。它支持一次移植目标应用程序的相同实现,然后在各种模拟器和真实网络环境中运行它,从而实现多样化和广泛的评估。我们确定了我们的方法的可行性,并且显示了可以忽略不计的内存和运行时开销。
{"title":"Towards a Common Interface for Overlay Network Simulators","authors":"C. Gross, Max Lehn, D. Stingl, A. Kovacevic, A. Buchmann, R. Steinmetz","doi":"10.1109/ICPADS.2010.33","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.33","url":null,"abstract":"Simulation has become an important evaluation method in the area of Peer-to-Peer (P2P) research due to the scalability limitations of evaluation test beds such as Planet Lab or G-Lab. Current simulators provide various abstraction levels for different underlay models, such that applications can be evaluated at different granularity. However, existing simulators suffer from a lack of interoperability and portability making the comparison of research results extremely difficult. To overcome this problem, we present an approach for a generic application interface for discrete-event P2P overlay network simulators. It enables porting of the same implementation of a targeted application once and then running it on various simulators as well as in a real network environment, thereby enabling a diverse and extensive evaluation. We established the feasibility of our approach and showed negligible memory and runtime overhead.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124405182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Data Vitalization: A New Paradigm for Large-Scale Dataset Analysis 数据活化:大规模数据集分析的新范式
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.102
Zhang Xiong, Wuman Luo, Lei Chen, L. Ni
Nowadays, datasets grow enormously both in size and complexity. One of the key issues confronted by large-scale dataset analysis is how to adapt systems to new, unprecedented query loads. Existing systems nail down the data organization scheme once and for all at the beginning of the system design, thus inevitably will see the performance goes down when user requirements change. In this paper, we propose a new paradigm, Data Vitalization, for large-scale dataset analysis. Our goal is to enable high flexibility such that the system is adaptive to complex analytical applications. Specifically, data are organized into a group of vitalized cells, each of which is a collection of data coupled with computing power. As user requirements change over time, cells evolve spontaneously to meet the potential new query loads. Besides basic functionality of Data Vitalization, we also explore an envisioned architecture of Data Vitalization including possible approaches for query processing, data evolution, as well as its tight-coupled mechanism for data storage and computing.
如今,数据集的规模和复杂性都在急剧增长。大规模数据集分析面临的关键问题之一是如何使系统适应新的、前所未有的查询负载。现有系统在设计之初就一劳永逸地确定了数据组织方案,因此当用户需求发生变化时,系统性能不可避免地会下降。在本文中,我们提出了一个新的范式,数据活化,用于大规模数据集分析。我们的目标是实现高灵活性,使系统能够适应复杂的分析应用。具体地说,数据被组织成一组活化的单元,每个单元都是与计算能力相结合的数据集合。随着时间的推移,用户需求会发生变化,计算单元会自发地进化,以满足潜在的新查询负载。除了数据赋能的基本功能外,我们还探讨了数据赋能的设想架构,包括查询处理、数据演化的可能方法,以及数据存储和计算的紧密耦合机制。
{"title":"Data Vitalization: A New Paradigm for Large-Scale Dataset Analysis","authors":"Zhang Xiong, Wuman Luo, Lei Chen, L. Ni","doi":"10.1109/ICPADS.2010.102","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.102","url":null,"abstract":"Nowadays, datasets grow enormously both in size and complexity. One of the key issues confronted by large-scale dataset analysis is how to adapt systems to new, unprecedented query loads. Existing systems nail down the data organization scheme once and for all at the beginning of the system design, thus inevitably will see the performance goes down when user requirements change. In this paper, we propose a new paradigm, Data Vitalization, for large-scale dataset analysis. Our goal is to enable high flexibility such that the system is adaptive to complex analytical applications. Specifically, data are organized into a group of vitalized cells, each of which is a collection of data coupled with computing power. As user requirements change over time, cells evolve spontaneously to meet the potential new query loads. Besides basic functionality of Data Vitalization, we also explore an envisioned architecture of Data Vitalization including possible approaches for query processing, data evolution, as well as its tight-coupled mechanism for data storage and computing.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121600685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Packet Content Matching with packetC Searchsets 数据包内容与packetC搜索集匹配
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.52
R. Duncan, P. Jungck, Kenneth Ross, S. Tillman
Increasing speeds and volumes push network packet applications to use parallel processing to boost performance. Examining the packet payload (message content) is a key aspect of packet processing. Applications search payloads to find strings that match a pattern described by regular expressions (regex). Searching for multiple strings that may start anywhere in the payload is a major obstacle to performance. Commercial systems often employ multiple network processors to provide parallel processing in general and use regex software engines or special regex processors to speed up searching performance via parallelism. Typically, regex rules are prepared separately from the application program and compiled into a binary image to be read by a regex processor or software engine. Our approach integrates specifying search rules with specifying network application code written in packet C, a C dialect that hides host-machine specifics, supports coarse-grain parallelism and supplies high-level data type and operator extensions for packet processing. packetC provides a search set data type, as well as match and find operations, to support payload searching. We show that our search set operator implementation, using associative memory and regex processors, lets users enjoy the performance benefits of parallel regex technology without learning hardware-specifics or using a separate regex toolchain’s use.
速度和容量的增加促使网络数据包应用程序使用并行处理来提高性能。检查包的有效负载(消息内容)是包处理的一个关键方面。应用程序搜索有效负载以查找与正则表达式(regex)描述的模式匹配的字符串。搜索可能从有效负载的任何位置开始的多个字符串是性能的主要障碍。商业系统通常使用多个网络处理器来提供并行处理,并使用regex软件引擎或特殊的regex处理器来通过并行性加快搜索性能。通常,regex规则与应用程序分开准备,并编译成二进制映像,供regex处理器或软件引擎读取。我们的方法集成了指定搜索规则和指定用包C编写的网络应用程序代码,包C是一种隐藏主机细节的C方言,支持粗粒度并行性,并为包处理提供高级数据类型和操作符扩展。packetC提供了一个搜索集数据类型,以及匹配和查找操作,以支持有效负载搜索。我们展示了使用关联内存和正则表达式处理器的搜索集运算符实现,使用户无需学习硬件细节或使用单独的正则表达式工具链就可以享受并行正则表达式技术的性能优势。
{"title":"Packet Content Matching with packetC Searchsets","authors":"R. Duncan, P. Jungck, Kenneth Ross, S. Tillman","doi":"10.1109/ICPADS.2010.52","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.52","url":null,"abstract":"Increasing speeds and volumes push network packet applications to use parallel processing to boost performance. Examining the packet payload (message content) is a key aspect of packet processing. Applications search payloads to find strings that match a pattern described by regular expressions (regex). Searching for multiple strings that may start anywhere in the payload is a major obstacle to performance. Commercial systems often employ multiple network processors to provide parallel processing in general and use regex software engines or special regex processors to speed up searching performance via parallelism. Typically, regex rules are prepared separately from the application program and compiled into a binary image to be read by a regex processor or software engine. Our approach integrates specifying search rules with specifying network application code written in packet C, a C dialect that hides host-machine specifics, supports coarse-grain parallelism and supplies high-level data type and operator extensions for packet processing. packetC provides a search set data type, as well as match and find operations, to support payload searching. We show that our search set operator implementation, using associative memory and regex processors, lets users enjoy the performance benefits of parallel regex technology without learning hardware-specifics or using a separate regex toolchain’s use.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116617579","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Coordinated Selective Rejuvenation for Distributed Services 分布式服务的协同选择性复兴
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.10
Guanhua Tian, Dan Meng
Service availability and QoS, in terms of customer affecting performance metrics, is crucial for service systems. However, the increasing complexity in distributed service systems introduce hidden space for software faults, which undermine system availability, leading to fault or even down time. In this paper, we introduce a composition technique, Coordinated Selective Rejuvenation, to automate the whole procession of fault component identification and rejuvenation arbitration, in order to guarantee distributed service system's customer-affecting metrics. We take evaluation with fault injection experiment on RUBiS, which simulates distributed eCommerce of eBay.com. The results indicate that our request path analysis approach and system model technique are effective for fault component's location, Bayesian network technique is feasible for fault pinpointing, in terms of request tracing context. Meanwhile, the arbitration scheme, can effectively guarantee system QoS, by identifying and rejuvenating most likely performance fault tier, before the degradation of customer affecting performance metric become severe.
就客户影响性能指标而言,服务可用性和QoS对服务系统至关重要。然而,随着分布式业务系统复杂性的不断提高,软件故障为其提供了隐藏空间,从而降低了系统的可用性,导致系统故障甚至停机。为了保证分布式服务系统的客户影响指标,本文引入了一种组合技术——协同选择恢复技术,实现故障组件识别和恢复仲裁的全过程自动化。采用故障注入实验对仿真ebay分布式电子商务的RUBiS进行了评价。结果表明,我们的请求路径分析方法和系统模型技术对故障组件的定位是有效的,贝叶斯网络技术在请求跟踪上下文方面对故障精确定位是可行的。同时,仲裁方案在客户降级影响性能指标变得严重之前,通过识别和恢复最有可能出现性能故障的层,有效地保证了系统的QoS。
{"title":"Coordinated Selective Rejuvenation for Distributed Services","authors":"Guanhua Tian, Dan Meng","doi":"10.1109/ICPADS.2010.10","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.10","url":null,"abstract":"Service availability and QoS, in terms of customer affecting performance metrics, is crucial for service systems. However, the increasing complexity in distributed service systems introduce hidden space for software faults, which undermine system availability, leading to fault or even down time. In this paper, we introduce a composition technique, Coordinated Selective Rejuvenation, to automate the whole procession of fault component identification and rejuvenation arbitration, in order to guarantee distributed service system's customer-affecting metrics. We take evaluation with fault injection experiment on RUBiS, which simulates distributed eCommerce of eBay.com. The results indicate that our request path analysis approach and system model technique are effective for fault component's location, Bayesian network technique is feasible for fault pinpointing, in terms of request tracing context. Meanwhile, the arbitration scheme, can effectively guarantee system QoS, by identifying and rejuvenating most likely performance fault tier, before the degradation of customer affecting performance metric become severe.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129596930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid Checkpointing for MPI Jobs in HPC Environments HPC环境中MPI作业的混合检查点
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.48
Chao Wang, F. Mueller, C. Engelmann, S. Scott
As the core count in high-performance computing systems keeps increasing, faults are becoming common place. Check pointing addresses such faults but captures full process images even though only a subset of the process image changes between checkpoints. We have designed a hybrid check pointing technique for MPI tasks of high-performance applications. This technique alternates between full and incremental checkpoints: At incremental checkpoints, only data changed since the last checkpoint is captured. Our implementation integrates new BLCR and LAM/MPI features that complement traditional full checkpoints. This results in significantly reduced checkpoint sizes and overheads with only moderate increases in restart overhead. After accounting for cost and savings, benefits due to incremental checkpoints are an order of magnitude larger than overheads on restarts. We further derive qualitative results indicating an optimal balance between full/incremental checkpoints of our novel approach at a ratio of 1:9, which outperforms both always-full and always-incremental check pointing.
随着高性能计算系统中核心数量的不断增加,故障也变得越来越普遍。检查点处理此类错误,但捕获完整的进程映像,即使在检查点之间只有进程映像的子集发生更改。我们为高性能应用程序的MPI任务设计了一种混合检查点技术。该技术在完整检查点和增量检查点之间交替使用:在增量检查点,只捕获自上次检查点以来更改的数据。我们的实现集成了新的BLCR和LAM/MPI功能,补充了传统的完整检查点。这将显著减少检查点大小和开销,而重启开销仅略有增加。在考虑了成本和节省之后,增量检查点带来的好处要比重新启动带来的开销大一个数量级。我们进一步得出定性结果,表明我们的新方法的完整/增量检查点之间以1:9的比例达到最佳平衡,其性能优于总是满的和总是增量的检查点。
{"title":"Hybrid Checkpointing for MPI Jobs in HPC Environments","authors":"Chao Wang, F. Mueller, C. Engelmann, S. Scott","doi":"10.1109/ICPADS.2010.48","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.48","url":null,"abstract":"As the core count in high-performance computing systems keeps increasing, faults are becoming common place. Check pointing addresses such faults but captures full process images even though only a subset of the process image changes between checkpoints. We have designed a hybrid check pointing technique for MPI tasks of high-performance applications. This technique alternates between full and incremental checkpoints: At incremental checkpoints, only data changed since the last checkpoint is captured. Our implementation integrates new BLCR and LAM/MPI features that complement traditional full checkpoints. This results in significantly reduced checkpoint sizes and overheads with only moderate increases in restart overhead. After accounting for cost and savings, benefits due to incremental checkpoints are an order of magnitude larger than overheads on restarts. We further derive qualitative results indicating an optimal balance between full/incremental checkpoints of our novel approach at a ratio of 1:9, which outperforms both always-full and always-incremental check pointing.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128932433","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 53
A Qualitative Analysis of Uncertainty and Correlation Computing for the Business Processes of Enterprise Interoperability 企业互操作性业务流程的不确定性定性分析与关联计算
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.100
Xiaofeng Liu, Xiaofei Xu, S. Deng
In the domain of enterprise interoperability many uncertain factors affect the performance of the whole cross-organizational business process, e.g. uncertain business process executing time, uncertain business logic in a process, et al. Uncertain factors couldn’t be avoided but can be analyzed. In this paper a model about Enterprise Interoperability Domain (EID) is given and the main uncertain factors during enterprise interoperability are analyzed. To analyze the correlation between the business processes in an EID, an updated grey correlation analysis method is given to help calculating the grey relational degree between the elements with uncertainty in an EID. The simulation shows that the result of grey correlation degree can be very helpful for the further optimization of enterprise interoperability business process.
在企业互操作性领域,许多不确定因素影响着整个跨组织业务流程的性能,如业务流程执行时间的不确定、流程中的业务逻辑的不确定等。不确定因素无法避免,但可以分析。本文建立了企业互操作域模型,分析了企业互操作过程中的主要不确定因素。为了分析EID中业务流程之间的关联,提出了一种改进的灰色关联分析方法,用于计算EID中不确定元素之间的灰色关联度。仿真结果表明,灰色关联度的计算结果对企业互操作性业务流程的进一步优化有很大的帮助。
{"title":"A Qualitative Analysis of Uncertainty and Correlation Computing for the Business Processes of Enterprise Interoperability","authors":"Xiaofeng Liu, Xiaofei Xu, S. Deng","doi":"10.1109/ICPADS.2010.100","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.100","url":null,"abstract":"In the domain of enterprise interoperability many uncertain factors affect the performance of the whole cross-organizational business process, e.g. uncertain business process executing time, uncertain business logic in a process, et al. Uncertain factors couldn’t be avoided but can be analyzed. In this paper a model about Enterprise Interoperability Domain (EID) is given and the main uncertain factors during enterprise interoperability are analyzed. To analyze the correlation between the business processes in an EID, an updated grey correlation analysis method is given to help calculating the grey relational degree between the elements with uncertainty in an EID. The simulation shows that the result of grey correlation degree can be very helpful for the further optimization of enterprise interoperability business process.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130153099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2010 IEEE 16th International Conference on Parallel and Distributed Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1