首页 > 最新文献

2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing最新文献

英文 中文
Self Protecting Data Sharing Using Generic Policies 使用通用策略的自我保护数据共享
Shiping Chen, Danan Thilakanathan, Donna Xu, S. Nepal, R. Calvo
Although content sharing provides many benefits, content owners lose full control of their content once they are given away. Existing solutions provide limited capabilities of content access control as they are vendor-specific, non-structured and non-flexible. In this paper, we present an open and flexible software solution called SelfProtect Object (SPO). SPO bundles content and policy files in an object that can protect its contents by itself anywhere and anytime. Our policy is based on XACML, a generic policy language allowing fine-grain access with rules and conditions. We also design and implement a prototype of SPO and demonstrate its capability through examples. Our solution is flexible to express a variety of access control rules and open to integrate into different applications on different platforms.
尽管内容共享提供了许多好处,但内容所有者一旦被分发出去,就失去了对其内容的完全控制。现有的解决方案提供的内容访问控制功能有限,因为它们是特定于供应商的、非结构化的和不灵活的。在本文中,我们提出了一个开放和灵活的软件解决方案,称为自我保护对象(SPO)。SPO将内容和策略文件捆绑在一个对象中,该对象可以随时随地保护其内容。我们的策略基于XACML,这是一种通用策略语言,允许使用规则和条件进行细粒度访问。我们还设计并实现了SPO的原型,并通过实例验证了其功能。我们的解决方案可以灵活地表达各种访问控制规则,并且可以开放地集成到不同平台上的不同应用程序中。
{"title":"Self Protecting Data Sharing Using Generic Policies","authors":"Shiping Chen, Danan Thilakanathan, Donna Xu, S. Nepal, R. Calvo","doi":"10.1109/CCGrid.2015.84","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.84","url":null,"abstract":"Although content sharing provides many benefits, content owners lose full control of their content once they are given away. Existing solutions provide limited capabilities of content access control as they are vendor-specific, non-structured and non-flexible. In this paper, we present an open and flexible software solution called SelfProtect Object (SPO). SPO bundles content and policy files in an object that can protect its contents by itself anywhere and anytime. Our policy is based on XACML, a generic policy language allowing fine-grain access with rules and conditions. We also design and implement a prototype of SPO and demonstrate its capability through examples. Our solution is flexible to express a variety of access control rules and open to integrate into different applications on different platforms.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"31 1","pages":"1197-1200"},"PeriodicalIF":0.0,"publicationDate":"2015-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75029535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
MIC-Tandem: Parallel X!Tandem Using MIC on Tandem Mass Spectrometry Based Proteomics Data MIC-Tandem:平行X!基于串联质谱的蛋白质组学数据串联使用MIC
Pinjie He, Kenli Li
The widespread use of mass spectrometry for protein identification has created an urgent demand for improving computational efficiency of matching mass spectrometry data to protein databases. With the rapid development of chip technology and parallel computing technique, such as multi-core processor, many-core coprocessor and cluster of multi-node, the speed and performance of the major mass spectral search engines are continuously improving. In recent ten years, X!Tandem as a popular and representative open-source program in searching mass spectral has extended several parallel versions and obtains considerable speedups. However, because these parallel strategies are mainly based on cluster of nodes, higher costs (e.g., charge of electricity and maintenance) is needed to get limited speedups. Fortunately, Intel Many Integrated Core (MIC) architecture and Graphics Processing Unit (GPU) are ideal for this problem. In this paper, we present and implement a parallel strategy to X!Tandem using MIC called MIC-Tandem, That shows excellent speedups on commodity hardware and produces the same results as the original program.
质谱法在蛋白质鉴定中的广泛应用,迫切需要提高质谱数据与蛋白质数据库匹配的计算效率。随着多核处理器、多核协处理器、多节点集群等芯片技术和并行计算技术的快速发展,各大质谱搜索引擎的速度和性能都在不断提高。近十年来,X!Tandem作为一个流行且具有代表性的质谱搜索开源程序,扩展了多个并行版本,并获得了相当大的速度提升。然而,由于这些并行策略主要基于节点集群,因此需要更高的成本(例如,电费和维护费用)来获得有限的速度。幸运的是,英特尔多集成核心(MIC)架构和图形处理单元(GPU)是解决这个问题的理想选择。在本文中,我们提出并实现了X!使用称为MIC-Tandem的MIC串联,它在商用硬件上显示出出色的加速,并产生与原始程序相同的结果。
{"title":"MIC-Tandem: Parallel X!Tandem Using MIC on Tandem Mass Spectrometry Based Proteomics Data","authors":"Pinjie He, Kenli Li","doi":"10.1109/CCGrid.2015.31","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.31","url":null,"abstract":"The widespread use of mass spectrometry for protein identification has created an urgent demand for improving computational efficiency of matching mass spectrometry data to protein databases. With the rapid development of chip technology and parallel computing technique, such as multi-core processor, many-core coprocessor and cluster of multi-node, the speed and performance of the major mass spectral search engines are continuously improving. In recent ten years, X!Tandem as a popular and representative open-source program in searching mass spectral has extended several parallel versions and obtains considerable speedups. However, because these parallel strategies are mainly based on cluster of nodes, higher costs (e.g., charge of electricity and maintenance) is needed to get limited speedups. Fortunately, Intel Many Integrated Core (MIC) architecture and Graphics Processing Unit (GPU) are ideal for this problem. In this paper, we present and implement a parallel strategy to X!Tandem using MIC called MIC-Tandem, That shows excellent speedups on commodity hardware and produces the same results as the original program.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"42 1","pages":"717-720"},"PeriodicalIF":0.0,"publicationDate":"2015-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90329808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Partition-Aware Routing to Improve Network Isolation in Infiniband Based Multi-tenant Clusters 分区感知路由改善Infiniband多租户集群中的网络隔离
Feroz Zahid, Ernst Gunnar Gran, Bartosz Bogdanski, Bjørn Dag Johnsen, T. Skeie
InfiniBand (IB) is a widely used network interconnect for modern high-performance computing systems. In large IB fabrics, isolation of nodes is provided through partitioning. The routing algorithm, however, is unaware of these partitions in the network, Traffic flows belonging to different partitions might share links inside the network fabric. This sharing of intermediate links creates interference, which is particularly critical to avoid in multi-tenant environments like a cloud. In such systems, each tenant should experience predictable network performance, unaffected by the workload of other tenants. In addition, using current routing schemes, routes crossing partition boundaries are considered when distributing routes onto links in the network, despite the fact that these routes will never be used. The result is degraded load-balancing. In this paper, we present a novel partition-aware fat-tree routing algorithm, pFTree. The pFTree algorithm utilizes several mechanisms to provide network-wide isolation of partitions belonging to different tenant groups. Given the available network resources, pFTree starts by isolating partitions at the physical link level, and then moves on to utilize virtual lanes, if needed. Our experiments and simulations show that pFTree is able to significantly reduce the affect of inter-partition interference without any additional functional overhead. Furthermore, pFTree also provides improved load-balancing over the de facto standard IB fat-tree routing algorithm.
IB (InfiniBand)是一种广泛应用于现代高性能计算系统的网络互连技术。在大型IB结构中,通过分区提供节点隔离。然而,路由算法不知道网络中的这些分区,属于不同分区的流量可能在网络结构中共享链接。这种中间链接的共享会产生干扰,在像云这样的多租户环境中避免这种干扰尤为重要。在这样的系统中,每个租户都应该体验到可预测的网络性能,不受其他租户工作负载的影响。此外,使用当前的路由方案,在将路由分配到网络中的链路时,会考虑跨越分区边界的路由,尽管这些路由永远不会被使用。其结果是负载均衡降级。本文提出了一种新的分区感知胖树路由算法pFTree。pFTree算法利用几种机制为属于不同租户组的分区提供网络范围的隔离。给定可用的网络资源,pFTree首先在物理链路级别隔离分区,然后在需要时继续利用虚拟通道。我们的实验和仿真表明,pFTree能够显著降低分区间干扰的影响,而没有任何额外的功能开销。此外,pFTree还在事实上的标准IB胖树路由算法上提供了改进的负载均衡。
{"title":"Partition-Aware Routing to Improve Network Isolation in Infiniband Based Multi-tenant Clusters","authors":"Feroz Zahid, Ernst Gunnar Gran, Bartosz Bogdanski, Bjørn Dag Johnsen, T. Skeie","doi":"10.1109/CCGrid.2015.96","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.96","url":null,"abstract":"InfiniBand (IB) is a widely used network interconnect for modern high-performance computing systems. In large IB fabrics, isolation of nodes is provided through partitioning. The routing algorithm, however, is unaware of these partitions in the network, Traffic flows belonging to different partitions might share links inside the network fabric. This sharing of intermediate links creates interference, which is particularly critical to avoid in multi-tenant environments like a cloud. In such systems, each tenant should experience predictable network performance, unaffected by the workload of other tenants. In addition, using current routing schemes, routes crossing partition boundaries are considered when distributing routes onto links in the network, despite the fact that these routes will never be used. The result is degraded load-balancing. In this paper, we present a novel partition-aware fat-tree routing algorithm, pFTree. The pFTree algorithm utilizes several mechanisms to provide network-wide isolation of partitions belonging to different tenant groups. Given the available network resources, pFTree starts by isolating partitions at the physical link level, and then moves on to utilize virtual lanes, if needed. Our experiments and simulations show that pFTree is able to significantly reduce the affect of inter-partition interference without any additional functional overhead. Furthermore, pFTree also provides improved load-balancing over the de facto standard IB fat-tree routing algorithm.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"28 1","pages":"189-198"},"PeriodicalIF":0.0,"publicationDate":"2015-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88473759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
F/M-CIP: Implementing Flash Memory Cache Using Conservative Insertion and Promotion F/M-CIP:使用保守插入和提升实现闪存缓存
Pub Date : 2015-05-04 DOI: 10.1109/CCGrid.2015.119
J. Yang, Q. Yang
Flash memory SSD has emerged as a promising storage media and fits naturally as a cache between the system RAM and the disk due to its performance/cost characteristics. Managing such an SSD cache is challenging and traditional cache replacements do not work well because of SSDs asymmetric read/write performances and wearing issues. This paper presents a new cache replacement algorithm referred to as F/M-CIP that accelerates disk I/O greatly. The idea is dividing the traditional LRU list into 4 parts: candidate-list, SSD-list, RAM-list and eviction-buffer-list. Upon a cache miss, the metadata of the missed block is conservatively inserted into the candidate-list but the data itself is not cached. The block in the candidate-list is then conservatively promoted to the RAM-list upon the k-th miss. At the bottom of the RAM-list, the eviction-buffer accumulates LRU blocks to be written into the SSD cache in batches to exploit the internal parallelism of SSD. The SSD-list is managed using a combination of regency and frequency replacement policies by means of conservative promotion upon hits. To quantitatively evaluate the performance of F/M-CIP, a prototype has been built on Linux kernel at the generic block layer. Experimental results on standard benchmarks and real world traces have shown that F/M-CIP accelerates disk I/O performance up to an order of magnitude compared to the traditional hard disk storage and up to a factor of 3 compared to the traditional SSD cache algorithm in terms of application execution time. Furthermore, F/M-CIP substantially reduces write operations to the SSD implying prolonged durability.
闪存SSD已经成为一种很有前途的存储介质,由于其性能/成本特性,它自然适合作为系统RAM和磁盘之间的缓存。管理这样的SSD缓存是具有挑战性的,并且由于SSD的非对称读写性能和磨损问题,传统的缓存替换不能很好地工作。本文提出了一种新的缓存替换算法F/M-CIP,它大大加快了磁盘I/O的速度。这个想法是将传统的LRU列表分为4部分:候选列表、ssd列表、ram列表和驱逐缓冲区列表。在缓存丢失时,丢失块的元数据被保守地插入到候选列表中,但数据本身不被缓存。然后,候选列表中的块在第k次丢失时被保守地提升到ram列表中。在ram列表的底部,驱逐缓冲区累积LRU块,以便批量写入SSD缓存中,以利用SSD的内部并行性。ssd列表使用摄政和频率替换策略的组合进行管理,通过在命中时进行保守促销。为了定量评价F/M-CIP的性能,在Linux内核的通用块层构建了一个原型。标准基准测试和现实世界跟踪的实验结果表明,与传统硬盘存储相比,F/M-CIP将磁盘I/O性能提高了一个数量级,在应用程序执行时间方面,与传统SSD缓存算法相比,F/M-CIP将磁盘I/O性能提高了3倍。此外,F/M-CIP大大减少了对SSD的写操作,这意味着延长了耐用性。
{"title":"F/M-CIP: Implementing Flash Memory Cache Using Conservative Insertion and Promotion","authors":"J. Yang, Q. Yang","doi":"10.1109/CCGrid.2015.119","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.119","url":null,"abstract":"Flash memory SSD has emerged as a promising storage media and fits naturally as a cache between the system RAM and the disk due to its performance/cost characteristics. Managing such an SSD cache is challenging and traditional cache replacements do not work well because of SSDs asymmetric read/write performances and wearing issues. This paper presents a new cache replacement algorithm referred to as F/M-CIP that accelerates disk I/O greatly. The idea is dividing the traditional LRU list into 4 parts: candidate-list, SSD-list, RAM-list and eviction-buffer-list. Upon a cache miss, the metadata of the missed block is conservatively inserted into the candidate-list but the data itself is not cached. The block in the candidate-list is then conservatively promoted to the RAM-list upon the k-th miss. At the bottom of the RAM-list, the eviction-buffer accumulates LRU blocks to be written into the SSD cache in batches to exploit the internal parallelism of SSD. The SSD-list is managed using a combination of regency and frequency replacement policies by means of conservative promotion upon hits. To quantitatively evaluate the performance of F/M-CIP, a prototype has been built on Linux kernel at the generic block layer. Experimental results on standard benchmarks and real world traces have shown that F/M-CIP accelerates disk I/O performance up to an order of magnitude compared to the traditional hard disk storage and up to a factor of 3 compared to the traditional SSD cache algorithm in terms of application execution time. Furthermore, F/M-CIP substantially reduces write operations to the SSD implying prolonged durability.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"256 1","pages":"81-90"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73554132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Towards Self Adaptable Security Monitoring in IaaS Clouds 迈向IaaS云中的自适应安全监控
Pub Date : 2015-05-04 DOI: 10.1109/CCGrid.2015.133
Anna Giannakou, Louis Rilling, Jean-Louis Pazat, Frédéric Majorczyk, C. Morin
Traditional intrusion detection systems are not adaptive enough to cope with the dynamic characteristics of cloud-hosted virtual infrastructures. This makes them unable to address new cloud-oriented security issues. In this paper we introduce SAIDS, a self-adaptable intrusion detection system tailored for cloud environments. SAIDS is designed to re-configure its components based on environmental changes. A prototype of SAIDS is described.
传统的入侵检测系统无法适应云托管虚拟基础设施的动态特性。这使得他们无法解决新的面向云的安全问题。本文介绍了一种针对云环境定制的自适应入侵检测系统SAIDS。SAIDS设计用于根据环境变化重新配置其组件。介绍了一种新型抗炎药的原型。
{"title":"Towards Self Adaptable Security Monitoring in IaaS Clouds","authors":"Anna Giannakou, Louis Rilling, Jean-Louis Pazat, Frédéric Majorczyk, C. Morin","doi":"10.1109/CCGrid.2015.133","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.133","url":null,"abstract":"Traditional intrusion detection systems are not adaptive enough to cope with the dynamic characteristics of cloud-hosted virtual infrastructures. This makes them unable to address new cloud-oriented security issues. In this paper we introduce SAIDS, a self-adaptable intrusion detection system tailored for cloud environments. SAIDS is designed to re-configure its components based on environmental changes. A prototype of SAIDS is described.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"81 1","pages":"737-740"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74173136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Predicting the Performance of Parallel Computing Models Using Queuing System 利用排队系统预测并行计算模型的性能
Chao Shen, W. Tong, Samina Kausar
Computing models provide the parallel and distributed algorithms for cloud. The ability to estimate the performance of parallel computing models for efficient resource scheduling is critical. Current techniques for predicting the performance are mostly based on analyzing and simulating. The behavior of parallel computing model directly leads to the diversity of mathematical model. Without a general prediction model, it is very hard to compare fairly different parallel computing models in several critical aspects, including computing capacity, resource configuration, scalability, fault tolerance and so on. In this paper, we design a mathematical model for predicting the performance by using queuing system. We make various computing models as a service system for shielding the diversity. The performance can be accurately estimated with the job waiting time and the job performing time. The heterogeneity of computing nodes may also be considered.
计算模型为云计算提供了并行和分布式算法。评估并行计算模型的性能以实现有效的资源调度的能力是至关重要的。目前的性能预测技术主要是基于分析和模拟。并行计算模型的特性直接导致了数学模型的多样性。如果没有通用的预测模型,就很难在几个关键方面比较相当不同的并行计算模型,包括计算能力、资源配置、可伸缩性、容错性等。在本文中,我们设计了一个用排队系统来预测性能的数学模型。我们制作了各种计算模型作为屏蔽分集的服务系统。利用作业等待时间和作业执行时间可以准确估计作业的性能。还可以考虑计算节点的异构性。
{"title":"Predicting the Performance of Parallel Computing Models Using Queuing System","authors":"Chao Shen, W. Tong, Samina Kausar","doi":"10.1109/CCGrid.2015.92","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.92","url":null,"abstract":"Computing models provide the parallel and distributed algorithms for cloud. The ability to estimate the performance of parallel computing models for efficient resource scheduling is critical. Current techniques for predicting the performance are mostly based on analyzing and simulating. The behavior of parallel computing model directly leads to the diversity of mathematical model. Without a general prediction model, it is very hard to compare fairly different parallel computing models in several critical aspects, including computing capacity, resource configuration, scalability, fault tolerance and so on. In this paper, we design a mathematical model for predicting the performance by using queuing system. We make various computing models as a service system for shielding the diversity. The performance can be accurately estimated with the job waiting time and the job performing time. The heterogeneity of computing nodes may also be considered.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"60 1","pages":"757-760"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75643744","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
ProRenaTa: Proactive and Reactive Tuning to Scale a Distributed Storage System ProRenaTa:扩展分布式存储系统的主动和被动调优
Y. Liu, N. Rameshan, Enric Monte-Moreno, Vladimir Vlassov, Leandro Navarro-Moldes
Provisioning tasteful services in the Cloud that guarantees high quality of service with reduced hosting cost is challenging to achieve. There are two typical auto-scaling approaches: predictive and reactive. A prediction based controller leaves the system enough time to react to workload changes while a feedback based controller scales the system with better accuracy. In this paper, we show the limitations of using a proactive or reactive approach in isolation to scale a tasteful system and the overhead involved. To overcome the limitations, we implement an elasticity controller, ProRenaTa, which combines both reactive and proactive approaches to leverage on their respective advantages and also implements a data migration model to handle the scaling overhead. We show that the combination of reactive and proactive approaches outperforms the state of the art approaches. Our experiments with Wikipedia workload trace indicate that ProRenaTa guarantees a high level of SLA commitments while improving the overall resource utilization.
在云中提供有品位的服务,以降低托管成本来保证高质量的服务,这是一项具有挑战性的任务。有两种典型的自动扩展方法:预测性和响应性。基于预测的控制器使系统有足够的时间对工作负载的变化作出反应,而基于反馈的控制器则以更好的精度扩展系统。在本文中,我们展示了单独使用主动或被动方法来扩展一个有品位的系统的局限性以及所涉及的开销。为了克服这些限制,我们实现了一个弹性控制器ProRenaTa,它结合了被动和主动方法来利用各自的优势,还实现了一个数据迁移模型来处理扩展开销。我们表明,反应性和主动性方法的结合优于最先进的方法。我们对Wikipedia工作负载跟踪的实验表明,ProRenaTa在提高整体资源利用率的同时保证了高水平的SLA承诺。
{"title":"ProRenaTa: Proactive and Reactive Tuning to Scale a Distributed Storage System","authors":"Y. Liu, N. Rameshan, Enric Monte-Moreno, Vladimir Vlassov, Leandro Navarro-Moldes","doi":"10.1109/CCGrid.2015.26","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.26","url":null,"abstract":"Provisioning tasteful services in the Cloud that guarantees high quality of service with reduced hosting cost is challenging to achieve. There are two typical auto-scaling approaches: predictive and reactive. A prediction based controller leaves the system enough time to react to workload changes while a feedback based controller scales the system with better accuracy. In this paper, we show the limitations of using a proactive or reactive approach in isolation to scale a tasteful system and the overhead involved. To overcome the limitations, we implement an elasticity controller, ProRenaTa, which combines both reactive and proactive approaches to leverage on their respective advantages and also implements a data migration model to handle the scaling overhead. We show that the combination of reactive and proactive approaches outperforms the state of the art approaches. Our experiments with Wikipedia workload trace indicate that ProRenaTa guarantees a high level of SLA commitments while improving the overall resource utilization.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"108 1","pages":"453-464"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75877063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Reconfigurations for Processor Arrays with Faulty Switches and Links 具有故障开关和链路的处理器阵列的重新配置
W. Jigang, Longting Zhu, Peilan He, Guiyuan Jiang
Large scale multiprocessor array suffers from frequent hardware defects or soft faults due to overheating, overload or occupancy by other running applications. To obtain fault-free logical array, reconfiguration techniques are proposed to reuse the fault-free PEs by changing the interconnection among PEs. Previous research has worked on this topic but assume that switches and links are fault-free. In this paper, we consider faults not only on the processing elements (PEs) but also on the switches and links, and develop efficient algorithms to construct as large as possible logical arrays with optimized networks length. To deal with the faults on switches and links, an efficient pre-processing procedure is designed, in which switch faults are transformed into link faults, and then faulty links are classified into several categories to handle. Then, we propose an efficient algorithm, A-MLA, to produce as many as possible logical columns which are then combined to form a two dimensional processor array. After that, we propose an algorithm A-TMLA to reduce the interconnection length of the logical array obtained by algorithm A-MLA, as short interconnect leads to small communication latency and power consumption. Extensive experimental results show that, even with switch faults and link faults, our approach can produce larger logical fault-free arrays with shorter interconnection length, compared to the state-of-the-art.
大型多处理器阵列由于过热、过载或被其他正在运行的应用程序占用而经常出现硬件缺陷或软故障。为了获得无故障逻辑阵列,提出了重构技术,通过改变pe之间的互连来重用无故障pe。先前的研究已经在这个主题上进行了工作,但假设交换机和链路是无故障的。在本文中,我们不仅考虑了处理单元的故障,而且考虑了交换机和链路的故障,并开发了有效的算法来构建具有优化网络长度的尽可能大的逻辑阵列。针对交换机和链路上的故障,设计了一种高效的预处理流程,将交换机故障转化为链路故障,然后将故障链路分类处理。然后,我们提出了一种高效的算法,a - mla,以产生尽可能多的逻辑列,然后组合成一个二维处理器阵列。之后,我们提出了一种算法A-TMLA,以减少算法A-MLA获得的逻辑阵列的互连长度,因为短的互连导致较小的通信延迟和功耗。广泛的实验结果表明,即使有开关故障和链路故障,与最先进的技术相比,我们的方法可以产生更大的逻辑无故障阵列,互连长度更短。
{"title":"Reconfigurations for Processor Arrays with Faulty Switches and Links","authors":"W. Jigang, Longting Zhu, Peilan He, Guiyuan Jiang","doi":"10.1109/CCGrid.2015.47","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.47","url":null,"abstract":"Large scale multiprocessor array suffers from frequent hardware defects or soft faults due to overheating, overload or occupancy by other running applications. To obtain fault-free logical array, reconfiguration techniques are proposed to reuse the fault-free PEs by changing the interconnection among PEs. Previous research has worked on this topic but assume that switches and links are fault-free. In this paper, we consider faults not only on the processing elements (PEs) but also on the switches and links, and develop efficient algorithms to construct as large as possible logical arrays with optimized networks length. To deal with the faults on switches and links, an efficient pre-processing procedure is designed, in which switch faults are transformed into link faults, and then faulty links are classified into several categories to handle. Then, we propose an efficient algorithm, A-MLA, to produce as many as possible logical columns which are then combined to form a two dimensional processor array. After that, we propose an algorithm A-TMLA to reduce the interconnection length of the logical array obtained by algorithm A-MLA, as short interconnect leads to small communication latency and power consumption. Extensive experimental results show that, even with switch faults and link faults, our approach can produce larger logical fault-free arrays with shorter interconnection length, compared to the state-of-the-art.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"86 1","pages":"141-148"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74536934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
YARNsim: Simulating Hadoop YARN YARNsim:模拟Hadoop YARN
Ning Liu, Xi Yang, Xian-He Sun, John Jenkins, R. Ross
Despite the popularity of the Apache Hadoop system, its success has been limited by issues such as single points of failure, centralized job/task management, and lack of support for programming models other than MapReduce. The next generation of Hadoop, Apache Hadoop YARN, is designed to address these issues. In this paper, we propose YARNsim, a simulation system for Hadoop YARN. YARNsim is based on parallel discrete event simulation and provides protocol-level accuracy in simulating key components of YARN. YARNsim provides a virtual platform on which system architects can evaluate the design and implementation of Hadoop YARN systems. Also, application developers can tune job performance and understand the tradeoffs between different configurations, and Hadoop YARN system vendors can evaluate system efficiency under limited budgets. To demonstrate the validity of YARNsim, we use it to model two real systems and compare the experimental results from YARNsim and the real systems. The experiments include standard Hadoop benchmarks, synthetic workloads, and a bioinformatics application. The results show that the error rate is within 10% for the majority of test cases. The experiments prove that YARNsim can provide what-if analysis for system designers in a timely manner and at minimal cost compared with testing and evaluating on a real system.
尽管Apache Hadoop系统很受欢迎,但它的成功受到了一些问题的限制,比如单点故障、集中式作业/任务管理,以及缺乏对MapReduce以外的编程模型的支持。下一代Hadoop, Apache Hadoop YARN,旨在解决这些问题。本文提出了基于Hadoop YARN的仿真系统YARNsim。YARNsim基于并行离散事件仿真,并在模拟YARN的关键组件时提供协议级精度。YARNsim提供了一个虚拟平台,系统架构师可以在上面评估Hadoop YARN系统的设计和实现。此外,应用程序开发人员可以调整作业性能并了解不同配置之间的权衡,Hadoop YARN系统供应商可以在有限的预算下评估系统效率。为了验证YARNsim的有效性,我们使用它对两个实际系统进行了建模,并将YARNsim的实验结果与实际系统进行了比较。实验包括标准Hadoop基准测试、合成工作负载和一个生物信息学应用程序。结果表明,大多数测试用例的错误率在10%以内。实验证明,与在真实系统上进行测试和评估相比,YARNsim能够以最小的成本及时为系统设计者提供what-if分析。
{"title":"YARNsim: Simulating Hadoop YARN","authors":"Ning Liu, Xi Yang, Xian-He Sun, John Jenkins, R. Ross","doi":"10.1109/CCGrid.2015.61","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.61","url":null,"abstract":"Despite the popularity of the Apache Hadoop system, its success has been limited by issues such as single points of failure, centralized job/task management, and lack of support for programming models other than MapReduce. The next generation of Hadoop, Apache Hadoop YARN, is designed to address these issues. In this paper, we propose YARNsim, a simulation system for Hadoop YARN. YARNsim is based on parallel discrete event simulation and provides protocol-level accuracy in simulating key components of YARN. YARNsim provides a virtual platform on which system architects can evaluate the design and implementation of Hadoop YARN systems. Also, application developers can tune job performance and understand the tradeoffs between different configurations, and Hadoop YARN system vendors can evaluate system efficiency under limited budgets. To demonstrate the validity of YARNsim, we use it to model two real systems and compare the experimental results from YARNsim and the real systems. The experiments include standard Hadoop benchmarks, synthetic workloads, and a bioinformatics application. The results show that the error rate is within 10% for the majority of test cases. The experiments prove that YARNsim can provide what-if analysis for system designers in a timely manner and at minimal cost compared with testing and evaluating on a real system.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"21 1","pages":"637-646"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73403525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Parallel Solving Method of SOR Based on the Numerical Marine Forecasting Model 基于数值海洋预报模型的SOR并行求解方法
Pub Date : 2015-05-04 DOI: 10.1109/CCGrid.2015.117
Renbo Pang, Jianliang Xu, Yunquan Zhang
The successive over-relaxation (SOR) algorithm is an important method to solve the linear equations in the numerical marine forecasting model, but it is not appropriate for parallelism. In this paper, the red-black ordering and the method to avoid cost of communication are used to implement the parallelism of the SOR algorithm and improve the performance of it. The experiments show that the performance of the parallel SOR algorithm with the red-black ordering and communication optimization is high, but the errors between the serial SOR algorithm and the parallel SOR algorithm are bigger and bigger with the increase of computing time steps. Based on the characteristics of the numerical marine forecasting model, a four-step parallel SOR algorithm is designed to solve the error problem.
逐次过松弛(SOR)算法是求解海洋数值预报模型中线性方程的重要方法,但不适合求解并行问题。本文采用红黑排序和避免通信代价的方法来实现SOR算法的并行性,提高算法的性能。实验表明,采用红黑排序和通信优化的并行SOR算法性能较高,但随着计算时间步长的增加,串行SOR算法与并行SOR算法之间的误差越来越大。针对海洋数值预报模型的特点,设计了一种四步并行SOR算法来解决误差问题。
{"title":"Parallel Solving Method of SOR Based on the Numerical Marine Forecasting Model","authors":"Renbo Pang, Jianliang Xu, Yunquan Zhang","doi":"10.1109/CCGrid.2015.117","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.117","url":null,"abstract":"The successive over-relaxation (SOR) algorithm is an important method to solve the linear equations in the numerical marine forecasting model, but it is not appropriate for parallelism. In this paper, the red-black ordering and the method to avoid cost of communication are used to implement the parallelism of the SOR algorithm and improve the performance of it. The experiments show that the performance of the parallel SOR algorithm with the red-black ordering and communication optimization is high, but the errors between the serial SOR algorithm and the parallel SOR algorithm are bigger and bigger with the increase of computing time steps. Based on the characteristics of the numerical marine forecasting model, a four-step parallel SOR algorithm is designed to solve the error problem.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"24 1","pages":"733-736"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75323330","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1