首页 > 最新文献

2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing最新文献

英文 中文
Spy: A QoS-Aware Anonymous Multi-Cloud Storage System Supporting DSSE Spy:一个支持DSSE的qos感知匿名多云存储系统
Pengyan Shen, Kai Guo, Mingzhong Xiao, Quanqing Xu
Constructing an overlay storage system based on multiple personal cloud storages is a desirable technique and novel idea for cloud storages. Existing designs provide the basic functions with some customized features. Unfortunately, some important issues have always been ignored including privacy protection, QoS and cipher-text search. In this paper, we present Spy, our design for an anonymous storage overlay network on multiple personal cloud storage, supporting a flexible QoS awareness and cipher-text search. We reform the original Tor protocol by extending the command set and adding a tail part to the Tor cell, which makes it possible for coordination among proxy servers and still keeps the anonymity. Based on which, we proposed a flexible user-defined QoS policy and employed a Dynamic Searchable Symmetric Encryption (DSSE) scheme to support secure cipher-text search. Extensive security analysis prove the security on privacy preserving and experiments show how different QoS policy work according to different security requirements.
构建基于多个个人云存储的覆盖存储系统是云存储的理想技术和新思路。现有的设计提供了基本的功能和一些定制的特性。不幸的是,一些重要的问题一直被忽视,包括隐私保护、QoS和密文搜索。在本文中,我们介绍了Spy,这是我们在多个个人云存储上设计的匿名存储覆盖网络,支持灵活的QoS感知和密文搜索。我们对原有的Tor协议进行了改进,扩展了命令集,并在Tor单元中增加了尾部部分,使代理服务器之间的协调成为可能,同时保持了匿名性。在此基础上,提出了灵活的用户自定义QoS策略,并采用动态可搜索对称加密(DSSE)方案支持安全密文搜索。大量的安全性分析证明了QoS策略在隐私保护方面的安全性,并通过实验证明了不同的QoS策略是如何根据不同的安全需求发挥作用的。
{"title":"Spy: A QoS-Aware Anonymous Multi-Cloud Storage System Supporting DSSE","authors":"Pengyan Shen, Kai Guo, Mingzhong Xiao, Quanqing Xu","doi":"10.1109/CCGrid.2015.88","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.88","url":null,"abstract":"Constructing an overlay storage system based on multiple personal cloud storages is a desirable technique and novel idea for cloud storages. Existing designs provide the basic functions with some customized features. Unfortunately, some important issues have always been ignored including privacy protection, QoS and cipher-text search. In this paper, we present Spy, our design for an anonymous storage overlay network on multiple personal cloud storage, supporting a flexible QoS awareness and cipher-text search. We reform the original Tor protocol by extending the command set and adding a tail part to the Tor cell, which makes it possible for coordination among proxy servers and still keeps the anonymity. Based on which, we proposed a flexible user-defined QoS policy and employed a Dynamic Searchable Symmetric Encryption (DSSE) scheme to support secure cipher-text search. Extensive security analysis prove the security on privacy preserving and experiments show how different QoS policy work according to different security requirements.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"17 1","pages":"951-960"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75318429","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Analysing Cancer Genomics in the Elastic Cloud 在弹性云中分析癌症基因组学
Pub Date : 2015-05-04 DOI: 10.1109/CCGRID.2015.176
Christopher Smowton, Crispin J. Miller, W. Xing, Andoena Balla, D. Antoniades, G. Pallis, M. Dikaiakos
With the rapidly growing demand for DNA analysis, the need for storing and processing large-scale genome data has presented significant challenges. This paper describes how the Genome Analysis Toolkit (GATK) can be deployed to an elastic cloud, and defines policy to drive elastic scaling of the application. We extensively analyse the GATK to expose opportunities for resource elasticity, demonstrate that it can be practically deployed at scale in a cloud environment, and demonstrate that applying elastic scaling improves the performance to cost tradeoff achieved in a simulated environment.
随着DNA分析需求的快速增长,存储和处理大规模基因组数据的需求提出了重大挑战。本文描述了如何将Genome Analysis Toolkit (GATK)部署到弹性云中,并定义了驱动应用程序弹性扩展的策略。我们对GATK进行了广泛的分析,以揭示资源弹性的机会,证明它可以在云环境中大规模部署,并证明应用弹性扩展可以提高性能,从而在模拟环境中实现成本权衡。
{"title":"Analysing Cancer Genomics in the Elastic Cloud","authors":"Christopher Smowton, Crispin J. Miller, W. Xing, Andoena Balla, D. Antoniades, G. Pallis, M. Dikaiakos","doi":"10.1109/CCGRID.2015.176","DOIUrl":"https://doi.org/10.1109/CCGRID.2015.176","url":null,"abstract":"With the rapidly growing demand for DNA analysis, the need for storing and processing large-scale genome data has presented significant challenges. This paper describes how the Genome Analysis Toolkit (GATK) can be deployed to an elastic cloud, and defines policy to drive elastic scaling of the application. We extensively analyse the GATK to expose opportunities for resource elasticity, demonstrate that it can be practically deployed at scale in a cloud environment, and demonstrate that applying elastic scaling improves the performance to cost tradeoff achieved in a simulated environment.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"80 1","pages":"835-844"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75905599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Improving Application Performance by Efficiently Utilizing Heterogeneous Many-core Platforms 有效利用异构多核平台提高应用程序性能
Jie Shen, A. Varbanescu, H. Sips
Heterogeneous platforms integrating different types of processing units (such as multi-core CPUs and GPUs) are in high demand in high performance computing. Existing studies have shown that using heterogeneous platforms can improve application performance and hardware utilization. However, systematic methods to design, implement, and map applications to efficiently use heterogeneous computing resources are only very few. The goal of my PhD research is therefore to study such heterogeneous systems and propose systematic methods to allow many (classes of) applications to efficiently use them. After 3.5 years of PhD study, my contributions are (1) a thorough evaluation of a suitable programming model for heterogeneous computing, (2) a workload partitioning framework to accelerate parallel applications on heterogeneous platforms, (3) a modelling-based prediction method to determine the optimal workload partitioning, (4) a systematic approach to decide the best mapping between the application and the platform by choosing the best performing hardware configuration (Only-CPU, Only-GPU, or CPU+GPU with the workload partitioning). In the near future, I plan to apply my approach to large-scale applications and platforms to expand its usability and applicability.
高性能计算对集成不同类型处理单元(如多核cpu和gpu)的异构平台有很高的需求。已有研究表明,使用异构平台可以提高应用程序性能和硬件利用率。然而,系统地设计、实现和映射应用程序以有效地使用异构计算资源的方法很少。因此,我博士研究的目标是研究这种异构系统,并提出系统的方法,以允许许多(类)应用程序有效地使用它们。经过三年半的博士学习,我的贡献是:(1)对适合异构计算的编程模型进行了全面的评估,(2)在异构平台上加速并行应用程序的工作负载分区框架,(3)基于建模的预测方法来确定最佳工作负载分区,(4)通过选择性能最佳的硬件配置(Only-CPU, Only-GPU,或者CPU+GPU的工作负载分区)。在不久的将来,我计划将我的方法应用于大型应用程序和平台,以扩大其可用性和适用性。
{"title":"Improving Application Performance by Efficiently Utilizing Heterogeneous Many-core Platforms","authors":"Jie Shen, A. Varbanescu, H. Sips","doi":"10.1109/CCGrid.2015.44","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.44","url":null,"abstract":"Heterogeneous platforms integrating different types of processing units (such as multi-core CPUs and GPUs) are in high demand in high performance computing. Existing studies have shown that using heterogeneous platforms can improve application performance and hardware utilization. However, systematic methods to design, implement, and map applications to efficiently use heterogeneous computing resources are only very few. The goal of my PhD research is therefore to study such heterogeneous systems and propose systematic methods to allow many (classes of) applications to efficiently use them. After 3.5 years of PhD study, my contributions are (1) a thorough evaluation of a suitable programming model for heterogeneous computing, (2) a workload partitioning framework to accelerate parallel applications on heterogeneous platforms, (3) a modelling-based prediction method to determine the optimal workload partitioning, (4) a systematic approach to decide the best mapping between the application and the platform by choosing the best performing hardware configuration (Only-CPU, Only-GPU, or CPU+GPU with the workload partitioning). In the near future, I plan to apply my approach to large-scale applications and platforms to expand its usability and applicability.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"23 1","pages":"709-712"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80099177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
GERBIL: MPI+YARN
Pub Date : 2015-05-04 DOI: 10.1109/CCGrid.2015.137
Luna Xu, Min Li, Ali Raza Butt
Emerging big data applications comprise rich multi-faceted workflows with both compute-intensive and data-intensive tasks, and intricate communication patterns. While MapReduce is an effective model for data-intensive tasks, the MPI programming model may be better suited for extracting high-performance for compute-intensive tasks. Researchers have recognized this need to employ specialized models for different phases of a workflow, e.g., performing computations using MPI followed by visualizations using MapReduce. However, extant multi-cluster approaches are inefficient as they entail data movement across clusters and porting across data formats. Consequently, there is a crucial need for disparate programming models to co-exist on the same set of resources. In this paper, we address the above issue by designing GERBIL, a framework for transparently co-hosting unmodified MPI applications alongside MapReduce applications on the same cluster. GERBIL exploits YARN as the model agnostic resource negotiator, and provides an easy-to-use interface to the users. GERBIL bridges the fundamental mismatch between YARN and MPI by designing an MPI-aware resource allocation mechanism. We also support five different optimizations: minimizing job wait time, achieving inter-process locality, achieving desired cluster utilization, minimizing network traffic, and minimizing job execution time, all in a multi-tenant environment. Our evaluation shows that GERBIL enables MPI executions with performance comparable to a native MPI setup, and improve compute-intensive applications performance by up to 133% when compared to the corresponding MapReduce-based versions.
新兴的大数据应用包括丰富的多方面的工作流,包括计算密集型和数据密集型任务,以及复杂的通信模式。虽然MapReduce是数据密集型任务的有效模型,但MPI编程模型可能更适合于为计算密集型任务提取高性能。研究人员已经认识到需要为工作流的不同阶段使用专门的模型,例如,使用MPI执行计算,然后使用MapReduce进行可视化。然而,现有的多集群方法效率低下,因为它们需要跨集群移动数据和跨数据格式移植。因此,非常需要不同的编程模型在同一组资源上共存。在本文中,我们通过设计GERBIL来解决上述问题,GERBIL是一个框架,用于透明地在同一集群上共同托管未修改的MPI应用程序和MapReduce应用程序。GERBIL利用YARN作为与模型无关的资源协商器,并为用户提供易于使用的界面。GERBIL通过设计一种MPI感知的资源分配机制,弥合了YARN和MPI之间的根本不匹配。我们还支持五种不同的优化:最小化作业等待时间、实现进程间局部性、实现所需的集群利用率、最小化网络流量和最小化作业执行时间,所有这些都在多租户环境中实现。我们的评估表明,GERBIL使MPI执行的性能与本机MPI设置相当,并且与相应的基于mapreduce的版本相比,计算密集型应用程序的性能提高了133%。
{"title":"GERBIL: MPI+YARN","authors":"Luna Xu, Min Li, Ali Raza Butt","doi":"10.1109/CCGrid.2015.137","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.137","url":null,"abstract":"Emerging big data applications comprise rich multi-faceted workflows with both compute-intensive and data-intensive tasks, and intricate communication patterns. While MapReduce is an effective model for data-intensive tasks, the MPI programming model may be better suited for extracting high-performance for compute-intensive tasks. Researchers have recognized this need to employ specialized models for different phases of a workflow, e.g., performing computations using MPI followed by visualizations using MapReduce. However, extant multi-cluster approaches are inefficient as they entail data movement across clusters and porting across data formats. Consequently, there is a crucial need for disparate programming models to co-exist on the same set of resources. In this paper, we address the above issue by designing GERBIL, a framework for transparently co-hosting unmodified MPI applications alongside MapReduce applications on the same cluster. GERBIL exploits YARN as the model agnostic resource negotiator, and provides an easy-to-use interface to the users. GERBIL bridges the fundamental mismatch between YARN and MPI by designing an MPI-aware resource allocation mechanism. We also support five different optimizations: minimizing job wait time, achieving inter-process locality, achieving desired cluster utilization, minimizing network traffic, and minimizing job execution time, all in a multi-tenant environment. Our evaluation shows that GERBIL enables MPI executions with performance comparable to a native MPI setup, and improve compute-intensive applications performance by up to 133% when compared to the corresponding MapReduce-based versions.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"36 1","pages":"627-636"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77363377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Design of a More Scalable Database System 一个更具可扩展性的数据库系统的设计
Hang Zhuang, Kun Lu, Changlong Li, Mingming Sun, Hang Chen, Xuehai Zhou
With the development of cloud computing and internet, e-Commerce, e-Business and corporate world revenue are increasing with high rate. These areas require scalable and consistent databases. NoSQL databases such as HBase has been proven to scalability and well performance on cloud computing platforms. However, the inevitable special data with few increment and frequent access leads to hotspot data and unbalanced accessing distribution between data storage servers. Due to their properties, these data often cannot be stored in multiple tables. Some storage nodes become the bottleneck of the distributed storage system, therefore, it becomes difficult to improve the performance by increasing the number of nodes which severely limits the scalability of the storage system. In order to make the performance of the cluster increases with the size of the cluster simultaneously, we devise a new distributed database storage framework to solve those issues mentioned above by changing the storage and read-write mode of the hotspot data. This structure guarantees that the hotspot data will not aggregate in the same storage node, as it guarantees that the data is not too hot in a single storage node. We implement the scalable database based on Apache HBase, which achieve almost double performance of throughput considering heavy read-write pressure situation only with double reading substites. Besides, heavy load node owing to hotspot data will no longer present in the new distributed database.
随着云计算和互联网的发展,电子商务、电子商务和企业收入正在高速增长。这些领域需要可伸缩且一致的数据库。HBase等NoSQL数据库在云计算平台上已经被证明具有良好的可扩展性和性能。然而,由于不可避免地存在增量少、访问频繁的特殊数据,导致数据存储服务器之间存在热点数据和不均衡的访问分布。由于它们的属性,这些数据通常不能存储在多个表中。一些存储节点成为分布式存储系统的瓶颈,通过增加节点数量来提高性能变得困难,严重限制了存储系统的可扩展性。为了使集群的性能随着集群规模的增加而同步增长,我们设计了一种新的分布式数据库存储框架,通过改变热点数据的存储和读写模式来解决上述问题。这种结构保证了热点数据不会聚集在同一个存储节点上,因为它保证了数据在单个存储节点上不会太热。我们实现了基于Apache HBase的可扩展数据库,考虑到读写压力大的情况下,仅使用双读替代,就实现了几乎两倍的吞吐量性能。此外,新的分布式数据库将不再存在热点数据导致的重载节点。
{"title":"Design of a More Scalable Database System","authors":"Hang Zhuang, Kun Lu, Changlong Li, Mingming Sun, Hang Chen, Xuehai Zhou","doi":"10.1109/CCGrid.2015.70","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.70","url":null,"abstract":"With the development of cloud computing and internet, e-Commerce, e-Business and corporate world revenue are increasing with high rate. These areas require scalable and consistent databases. NoSQL databases such as HBase has been proven to scalability and well performance on cloud computing platforms. However, the inevitable special data with few increment and frequent access leads to hotspot data and unbalanced accessing distribution between data storage servers. Due to their properties, these data often cannot be stored in multiple tables. Some storage nodes become the bottleneck of the distributed storage system, therefore, it becomes difficult to improve the performance by increasing the number of nodes which severely limits the scalability of the storage system. In order to make the performance of the cluster increases with the size of the cluster simultaneously, we devise a new distributed database storage framework to solve those issues mentioned above by changing the storage and read-write mode of the hotspot data. This structure guarantees that the hotspot data will not aggregate in the same storage node, as it guarantees that the data is not too hot in a single storage node. We implement the scalable database based on Apache HBase, which achieve almost double performance of throughput considering heavy read-write pressure situation only with double reading substites. Besides, heavy load node owing to hotspot data will no longer present in the new distributed database.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"18 1","pages":"1213-1216"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74338266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
An Empirical Performance Evaluation of GPU-Enabled Graph-Processing Systems gpu支持的图形处理系统的经验性能评估
Yong Guo, A. Varbanescu, A. Iosup, D. Epema
Graph processing is increasingly used in knowledge economies and in science, in advanced marketing, social networking, bioinformatics, etc. A number of graph-processing systems, including the GPU-enabled Medusa and Totem, have been developed recently. Understanding their performance is key to system selection, tuning, and improvement. Previous performance evaluation studies have been conducted for CPU-based graph-processing systems, such as Graph and GraphX. Unlike them, the performance of GPU-enabled systems is still not thoroughly evaluated and compared. To address this gap, we propose an empirical method for evaluating GPU-enabled graph-processing systems, which includes new performance metrics and a selection of new datasets and algorithms. By selecting 9 diverse graphs and 3 typical graph-processing algorithms, we conduct a comparative performance study of 3 GPU-enabled systems, Medusa, Totem, and MapGraph. We present the first comprehensive evaluation of GPU-enabled systems with results giving insight into raw processing power, performance breakdown into core components, scalability, and the impact on performance of system-specific optimization techniques and of the GPU generation. We present and discuss many findings that would benefit users and developers interested in GPU acceleration for graph processing.
图形处理越来越多地应用于知识经济和科学、高级市场营销、社交网络、生物信息学等领域。最近开发了许多图形处理系统,包括支持gpu的Medusa和Totem。了解它们的性能是系统选择、调优和改进的关键。以前的性能评估研究是针对基于cpu的图形处理系统(如Graph和GraphX)进行的。与它们不同的是,支持gpu的系统的性能仍然没有得到彻底的评估和比较。为了解决这一差距,我们提出了一种评估gpu支持的图形处理系统的经验方法,其中包括新的性能指标和新数据集和算法的选择。通过选择9种不同的图形和3种典型的图形处理算法,我们对3种支持gpu的系统(Medusa、Totem和MapGraph)进行了性能比较研究。我们首次对支持GPU的系统进行了全面评估,结果深入了解了原始处理能力、核心组件的性能分解、可扩展性以及系统特定优化技术和GPU生成对性能的影响。我们提出并讨论了许多发现,这些发现将使对图形处理GPU加速感兴趣的用户和开发人员受益。
{"title":"An Empirical Performance Evaluation of GPU-Enabled Graph-Processing Systems","authors":"Yong Guo, A. Varbanescu, A. Iosup, D. Epema","doi":"10.1109/CCGrid.2015.20","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.20","url":null,"abstract":"Graph processing is increasingly used in knowledge economies and in science, in advanced marketing, social networking, bioinformatics, etc. A number of graph-processing systems, including the GPU-enabled Medusa and Totem, have been developed recently. Understanding their performance is key to system selection, tuning, and improvement. Previous performance evaluation studies have been conducted for CPU-based graph-processing systems, such as Graph and GraphX. Unlike them, the performance of GPU-enabled systems is still not thoroughly evaluated and compared. To address this gap, we propose an empirical method for evaluating GPU-enabled graph-processing systems, which includes new performance metrics and a selection of new datasets and algorithms. By selecting 9 diverse graphs and 3 typical graph-processing algorithms, we conduct a comparative performance study of 3 GPU-enabled systems, Medusa, Totem, and MapGraph. We present the first comprehensive evaluation of GPU-enabled systems with results giving insight into raw processing power, performance breakdown into core components, scalability, and the impact on performance of system-specific optimization techniques and of the GPU generation. We present and discuss many findings that would benefit users and developers interested in GPU acceleration for graph processing.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"75 1","pages":"423-432"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86352028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
On the Design of a Demo for Exhibiting rCUDA rCUDA展示演示的设计
C. Reaño, Ferran Perez, F. Silla
CUDA is a technology developed by NVIDIA which provides a parallel computing platform and programming model for NVIDIA GPUs and compatible ones. It takes benefit from the enormous parallel processing power of GPUs in order to accelerate a wide range of applications, thus reducing their execution time. rCUDA (remote CUDA) is a middleware which grants applications concurrent access to CUDA-compatible devices installed in other nodes of the cluster in a transparent way so that applications are not aware of being accessing a remote device. In this paper we present a demo which shows, in real time, the overhead introduced by rCUDA in comparison to CUDA when running image filtering applications. The approach followed in this work is to develop a graphical demo which contains both an appealing design and technical contents.
CUDA是NVIDIA开发的一项技术,为NVIDIA gpu及兼容gpu提供并行计算平台和编程模型。它利用gpu巨大的并行处理能力来加速各种应用程序,从而减少它们的执行时间。rCUDA(远程CUDA)是一种中间件,它允许应用程序以透明的方式并发访问安装在集群其他节点上的CUDA兼容设备,以便应用程序不知道正在访问远程设备。在本文中,我们提供了一个演示,实时显示了rCUDA与CUDA在运行图像过滤应用程序时引入的开销。在这项工作中遵循的方法是开发一个图形演示,其中既包含吸引人的设计又包含技术内容。
{"title":"On the Design of a Demo for Exhibiting rCUDA","authors":"C. Reaño, Ferran Perez, F. Silla","doi":"10.1109/CCGrid.2015.53","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.53","url":null,"abstract":"CUDA is a technology developed by NVIDIA which provides a parallel computing platform and programming model for NVIDIA GPUs and compatible ones. It takes benefit from the enormous parallel processing power of GPUs in order to accelerate a wide range of applications, thus reducing their execution time. rCUDA (remote CUDA) is a middleware which grants applications concurrent access to CUDA-compatible devices installed in other nodes of the cluster in a transparent way so that applications are not aware of being accessing a remote device. In this paper we present a demo which shows, in real time, the overhead introduced by rCUDA in comparison to CUDA when running image filtering applications. The approach followed in this work is to develop a graphical demo which contains both an appealing design and technical contents.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"35 1","pages":"1169-1172"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85086134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Augmenting Performance For Distributed Cloud Storage 增强分布式云存储的性能
Pub Date : 2015-05-04 DOI: 10.1109/CCGrid.2015.124
Matthew B. Hancock, Carlos A. Varela
The device people use to capture multimedia has changed over the years with the rise of smart phones. Smart phones are readily available, easy to use, and capture multimedia with high quality. While consumers capture all of this media, the storage requirements are not changing significantly. Therefore, people look towards cloud storage solutions. The typical consumer stores files within a single provider. They want a solution that is quick to access, reliable, and secure. Using multiple providers can reduce cost and improve overall performance. We present a middleware framework called Distributed Indexed Storage in the Cloud (DISC) to improve all aspects a user expects in a cloud provider. The process of uploading and downloading is essentially transparent to the user. The upload and download performance happens simultaneously by distributing a subset of the file across multiple cloud providers that it deems fit based on policies. Reliability is another important feature of DISC. To improve reliability, we propose a solution that replicates the same subset of the file across different providers. This is beneficial when one provider is unresponsive, the data can be pulled from another provider with the same subset. Security has great importance when dealing with consumers data. We inherently gain security when improving reliability. Since the file is distributed using subsets, not one provider has the full file. In our experiment, performance improvements are observed when delivering and retrieving files compared to the standard approach. The results are promising, saving upwards of eight seconds in processing time. With the expansion of more cloud providers, the results are expected to improve.
随着智能手机的兴起,人们用来捕捉多媒体的设备已经发生了变化。智能手机随时可用,易于使用,并且可以捕获高质量的多媒体。当消费者获取所有这些媒体时,存储需求并没有显著变化。因此,人们开始关注云存储解决方案。典型的消费者将文件存储在单个提供者中。他们想要一个快速访问、可靠和安全的解决方案。使用多个提供者可以降低成本并提高整体性能。我们提出了一个中间件框架,称为云中的分布式索引存储(DISC),以改进用户对云提供商的所有期望。上传和下载的过程对用户来说基本上是透明的。上传和下载性能是通过将文件的一个子集分发到它认为适合的基于策略的多个云提供商来同时实现的。可靠性是DISC的另一个重要特征。为了提高可靠性,我们提出了一种跨不同提供者复制相同文件子集的解决方案。当一个提供程序没有响应时,这是有益的,数据可以从具有相同子集的另一个提供程序中提取。在处理消费者数据时,安全性非常重要。在提高可靠性的同时,我们本质上也获得了安全性。由于文件是使用子集分发的,因此没有一个提供程序拥有完整的文件。在我们的实验中,与标准方法相比,在交付和检索文件时可以观察到性能的改进。结果很有希望,可以节省8秒以上的处理时间。随着更多云提供商的扩张,结果有望得到改善。
{"title":"Augmenting Performance For Distributed Cloud Storage","authors":"Matthew B. Hancock, Carlos A. Varela","doi":"10.1109/CCGrid.2015.124","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.124","url":null,"abstract":"The device people use to capture multimedia has changed over the years with the rise of smart phones. Smart phones are readily available, easy to use, and capture multimedia with high quality. While consumers capture all of this media, the storage requirements are not changing significantly. Therefore, people look towards cloud storage solutions. The typical consumer stores files within a single provider. They want a solution that is quick to access, reliable, and secure. Using multiple providers can reduce cost and improve overall performance. We present a middleware framework called Distributed Indexed Storage in the Cloud (DISC) to improve all aspects a user expects in a cloud provider. The process of uploading and downloading is essentially transparent to the user. The upload and download performance happens simultaneously by distributing a subset of the file across multiple cloud providers that it deems fit based on policies. Reliability is another important feature of DISC. To improve reliability, we propose a solution that replicates the same subset of the file across different providers. This is beneficial when one provider is unresponsive, the data can be pulled from another provider with the same subset. Security has great importance when dealing with consumers data. We inherently gain security when improving reliability. Since the file is distributed using subsets, not one provider has the full file. In our experiment, performance improvements are observed when delivering and retrieving files compared to the standard approach. The results are promising, saving upwards of eight seconds in processing time. With the expansion of more cloud providers, the results are expected to improve.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"23 1","pages":"1189-1192"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91026285","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Scalable In-Memory Computing 可扩展内存计算
Pub Date : 2015-05-04 DOI: 10.1109/CCGrid.2015.106
Alexandru Uta, Andreea Sandu, S. Costache, T. Kielmann
Data-intensive scientific workflows are composed of many tasks that exhibit data precedence constraints leading to communication schemes expressed by means of intermediate files. In such scenarios, the storage layer is often a bottleneck, limiting overall application scalability, due to large volumes of data being generated during runtime at high I/O rates. To alleviate the storage pressure, applications take advantage of in-memory runtime distributed file systems that act as a fast, distributed cache, which greatly enhances I/O performance.In this paper, we present scalability results for MemFS, a distributed in-memory runtime file system. MemFS takes an opposite approach to data locality, by scattering all data among the nodes, leading to well balanced storage and network traffic, and thus making the system both highly per formant and scalable. Our results show that MemFS is platform independent, performing equally well on both private clusters and commercial clouds. On such platforms, running on up to 1024 cores, MemFS shows excellent horizontal scalability (using more nodes), while the vertical scalability (using more cores per node) is only limited by the network b and with. Further more, for this challenge we show how MemFS is able to scale elastically, at runtime, based on the application storage demands. In our experiments, we have successfully used up to 1TB memory when running a large instance of the Montage workflow.
数据密集型科学工作流由许多任务组成,这些任务表现出数据优先约束,导致通过中间文件表示的通信方案。在这种情况下,存储层通常是一个瓶颈,限制了整个应用程序的可伸缩性,因为在运行期间会以高I/O速率生成大量数据。为了减轻存储压力,应用程序利用内存中的运行时分布式文件系统作为快速的分布式缓存,这大大提高了I/O性能。在本文中,我们给出了MemFS(一个分布式内存运行时文件系统)的可伸缩性结果。MemFS采用与数据局部性相反的方法,将所有数据分散到节点之间,从而实现良好的存储和网络流量平衡,从而使系统具有高性能和可伸缩性。我们的结果表明,MemFS是平台无关的,在私有集群和商业云上的表现都一样好。在这样的平台上,MemFS最多可运行1024个内核,它显示出出色的水平可伸缩性(使用更多节点),而垂直可伸缩性(每个节点使用更多内核)仅受网络b和网络的限制。此外,对于这个挑战,我们将展示MemFS如何能够在运行时根据应用程序存储需求进行弹性扩展。在我们的实验中,当运行一个大型的蒙太奇工作流实例时,我们已经成功地使用了高达1TB的内存。
{"title":"Scalable In-Memory Computing","authors":"Alexandru Uta, Andreea Sandu, S. Costache, T. Kielmann","doi":"10.1109/CCGrid.2015.106","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.106","url":null,"abstract":"Data-intensive scientific workflows are composed of many tasks that exhibit data precedence constraints leading to communication schemes expressed by means of intermediate files. In such scenarios, the storage layer is often a bottleneck, limiting overall application scalability, due to large volumes of data being generated during runtime at high I/O rates. To alleviate the storage pressure, applications take advantage of in-memory runtime distributed file systems that act as a fast, distributed cache, which greatly enhances I/O performance.In this paper, we present scalability results for MemFS, a distributed in-memory runtime file system. MemFS takes an opposite approach to data locality, by scattering all data among the nodes, leading to well balanced storage and network traffic, and thus making the system both highly per formant and scalable. Our results show that MemFS is platform independent, performing equally well on both private clusters and commercial clouds. On such platforms, running on up to 1024 cores, MemFS shows excellent horizontal scalability (using more nodes), while the vertical scalability (using more cores per node) is only limited by the network b and with. Further more, for this challenge we show how MemFS is able to scale elastically, at runtime, based on the application storage demands. In our experiments, we have successfully used up to 1TB memory when running a large instance of the Montage workflow.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"10 1","pages":"805-810"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90155076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Majority Quorum Protocol Dedicated to General Threshold Schemes 专用于一般阈值方案的多数仲裁协议
T. J. R. Relaza, J. Jorda, A. Mzoughi
In this paper, we introduce a majority quorum system dedicated to p-m-n general threshold schemes where p, n and m are respectively the minimal number of chunks that provide some information (but not necessarily all) on the original data, the total number of nodes in which the chunks of an object are stored and the minimal number of nodes needed to retrieve the original data using this protocol. In other words, less than p chunks reveal absolutely no information about the original data and less than m chunks can't reconstruct the original data. The p-m-n general threshold schemes optimize the usage of storage resources by reducing the total size of data to write and ensure fault-tolerance up to (n - m) nodes failure. With such a data distribution, a specific value of m can be set to have a good trade off between resources utilization and fault-tolerance. The only drawback of such schemes is the lack of any consistency protocol. If fact, consistency protocols like classical majority quorum are based on full replication. To successfully read or write a data using the majority quorum protocol, an absolute majority of replicas must be read / written correctly. This condition ensures that any read and write operations will contain at least one common replica, which guarantees their consistency. However, when a threshold scheme is used, an adaptation is needed. In fact, classical majority quorum protocol can no longer ensure that m chunks will have the latest version when [n/2] + 1 <; m ≤ n. In this paper, we introduce a new majority quorum protocol dedicated to general threshold schemes. As for the classical majority quorum protocol, the complexity of the quorum size of our protocol is O(n) but the utilization of storage resources is greatly optimized.
在本文中,我们引入了一个专门用于p-m-n一般阈值方案的多数仲裁系统,其中p, n和m分别是提供原始数据的一些信息(但不一定是全部)的最小块数量,存储对象块的节点总数以及使用该协议检索原始数据所需的最小节点数量。换句话说,少于p个块完全没有透露原始数据的信息,少于m个块无法重建原始数据。p-m-n通用阈值方案通过减少写入数据的总大小来优化存储资源的使用,并保证高达(n -m)个节点故障的容错能力。对于这样的数据分布,可以设置一个特定的m值,以便在资源利用率和容错性之间取得良好的平衡。这种方案的唯一缺点是缺乏一致性协议。事实上,像经典多数仲裁这样的一致性协议是基于完全复制的。要使用多数仲裁协议成功读取或写入数据,必须正确读取/写入绝对多数副本。此条件确保任何读写操作都至少包含一个公共副本,从而保证它们的一致性。然而,当使用阈值方案时,需要进行自适应。事实上,当[n/2] + 1 <;时,经典多数仲裁协议已不能保证m块拥有最新版本。在本文中,我们引入了一种新的专门用于一般阈值方案的多数仲裁协议。对于经典的多数仲裁协议,我们的协议的仲裁大小复杂度为0 (n),但存储资源的利用率得到了极大的优化。
{"title":"Majority Quorum Protocol Dedicated to General Threshold Schemes","authors":"T. J. R. Relaza, J. Jorda, A. Mzoughi","doi":"10.1109/CCGrid.2015.76","DOIUrl":"https://doi.org/10.1109/CCGrid.2015.76","url":null,"abstract":"In this paper, we introduce a majority quorum system dedicated to p-m-n general threshold schemes where p, n and m are respectively the minimal number of chunks that provide some information (but not necessarily all) on the original data, the total number of nodes in which the chunks of an object are stored and the minimal number of nodes needed to retrieve the original data using this protocol. In other words, less than p chunks reveal absolutely no information about the original data and less than m chunks can't reconstruct the original data. The p-m-n general threshold schemes optimize the usage of storage resources by reducing the total size of data to write and ensure fault-tolerance up to (n - m) nodes failure. With such a data distribution, a specific value of m can be set to have a good trade off between resources utilization and fault-tolerance. The only drawback of such schemes is the lack of any consistency protocol. If fact, consistency protocols like classical majority quorum are based on full replication. To successfully read or write a data using the majority quorum protocol, an absolute majority of replicas must be read / written correctly. This condition ensures that any read and write operations will contain at least one common replica, which guarantees their consistency. However, when a threshold scheme is used, an adaptation is needed. In fact, classical majority quorum protocol can no longer ensure that m chunks will have the latest version when [n/2] + 1 <; m ≤ n. In this paper, we introduce a new majority quorum protocol dedicated to general threshold schemes. As for the classical majority quorum protocol, the complexity of the quorum size of our protocol is O(n) but the utilization of storage resources is greatly optimized.","PeriodicalId":6664,"journal":{"name":"2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing","volume":"30 1","pages":"785-788"},"PeriodicalIF":0.0,"publicationDate":"2015-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84730930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1