首页 > 最新文献

2012 IEEE Fifth International Conference on Cloud Computing最新文献

英文 中文
Scalable Transaction Management with Snapshot Isolation on Cloud Data Management Systems 云数据管理系统上具有快照隔离的可扩展事务管理
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.102
Vinit Padhye, A. Tripathi
We address the problem of building scalable transaction management mechanisms for multi-row transactions on key-value storage systems. We develop scalable techniques for transaction management utilizing the snapshot isolation (SI)model. Because the SI model can lead to non-serializable transaction executions, we investigate two conflict detection techniques for ensuring serializability under SI. To support scalability, we investigate system models and mechanisms in which the transaction management functions are decoupled from the storage system and integrated with the application-level processes. We present two system models and demonstrate their scalability under the scale-out paradigm of Cloud computing platforms. In the first system model, all transaction management functions are executed in a fully decentralized manner by the application processes. The second model is based on a hybrid approach in which the conflict detection techniques are implemented by a dedicated service. We perform a comparative evaluation of these models using the TPC-C benchmark and demonstrate their scalability.
我们解决了在键值存储系统上为多行事务构建可扩展事务管理机制的问题。我们利用快照隔离(SI)模型开发可扩展的事务管理技术。由于SI模型可能导致不可序列化的事务执行,因此我们研究了两种冲突检测技术,以确保SI下的可序列化性。为了支持可伸缩性,我们研究了系统模型和机制,在这些模型和机制中,事务管理功能与存储系统解耦,并与应用程序级流程集成。我们提出了两种系统模型,并展示了它们在云计算平台横向扩展范式下的可扩展性。在第一个系统模型中,所有事务管理功能都由应用程序流程以完全分散的方式执行。第二个模型基于混合方法,其中冲突检测技术由专用服务实现。我们使用TPC-C基准对这些模型进行了比较评估,并展示了它们的可扩展性。
{"title":"Scalable Transaction Management with Snapshot Isolation on Cloud Data Management Systems","authors":"Vinit Padhye, A. Tripathi","doi":"10.1109/CLOUD.2012.102","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.102","url":null,"abstract":"We address the problem of building scalable transaction management mechanisms for multi-row transactions on key-value storage systems. We develop scalable techniques for transaction management utilizing the snapshot isolation (SI)model. Because the SI model can lead to non-serializable transaction executions, we investigate two conflict detection techniques for ensuring serializability under SI. To support scalability, we investigate system models and mechanisms in which the transaction management functions are decoupled from the storage system and integrated with the application-level processes. We present two system models and demonstrate their scalability under the scale-out paradigm of Cloud computing platforms. In the first system model, all transaction management functions are executed in a fully decentralized manner by the application processes. The second model is based on a hybrid approach in which the conflict detection techniques are implemented by a dedicated service. We perform a comparative evaluation of these models using the TPC-C benchmark and demonstrate their scalability.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"206 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114985620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
COSBench: A Benchmark Tool for Cloud Object Storage Services COSBench:云对象存储服务的基准测试工具
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.52
Qing Zheng, Hao-peng Chen, Yaguang Wang, Jiangang Duan, Zhiteng Huang
With object storage services becoming increasingly accepted as replacements for traditional file or block systems, it is important to effectively measure the performance of these services. Thus people can compare different solutions or tune their systems for better performance. However, little has been reported on this specific topic as yet. To address this problem, we present COSBench (Cloud Object Storage Benchmark), a benchmark tool that we are currently working on in Intel for cloud object storage services. In addition, in this paper, we also share the results of the experiments we have performed so far.
随着对象存储服务作为传统文件或块系统的替代品被越来越多地接受,有效地度量这些服务的性能变得非常重要。因此,人们可以比较不同的解决方案或调整他们的系统以获得更好的性能。然而,迄今为止关于这一具体主题的报道很少。为了解决这个问题,我们提出了COSBench(云对象存储基准),这是一个我们目前正在英特尔为云对象存储服务开发的基准测试工具。此外,在本文中,我们还分享了我们目前所做的实验结果。
{"title":"COSBench: A Benchmark Tool for Cloud Object Storage Services","authors":"Qing Zheng, Hao-peng Chen, Yaguang Wang, Jiangang Duan, Zhiteng Huang","doi":"10.1109/CLOUD.2012.52","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.52","url":null,"abstract":"With object storage services becoming increasingly accepted as replacements for traditional file or block systems, it is important to effectively measure the performance of these services. Thus people can compare different solutions or tune their systems for better performance. However, little has been reported on this specific topic as yet. To address this problem, we present COSBench (Cloud Object Storage Benchmark), a benchmark tool that we are currently working on in Intel for cloud object storage services. In addition, in this paper, we also share the results of the experiments we have performed so far.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"280 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121821782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Introducing STRATOS: A Cloud Broker Service 介绍STRATOS:云代理服务
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.24
P. Pawluk, B. Simmons, Michael Smit, Marin Litoiu, Serge Mankovskii
This paper introduces a cloud broker service (STRATOS) which facilitates the deployment and runtime management of cloud application topologies using cloud elements/services sourced on the fly from multiple providers, based on requirements specified in higher level objectives. Its implementation and use is evaluated in a set of experiments.
本文介绍了一个云代理服务(STRATOS),它可以根据更高级别目标中指定的需求,使用来自多个提供商的动态云元素/服务,促进云应用程序拓扑的部署和运行时管理。在一组实验中评估了它的实现和使用。
{"title":"Introducing STRATOS: A Cloud Broker Service","authors":"P. Pawluk, B. Simmons, Michael Smit, Marin Litoiu, Serge Mankovskii","doi":"10.1109/CLOUD.2012.24","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.24","url":null,"abstract":"This paper introduces a cloud broker service (STRATOS) which facilitates the deployment and runtime management of cloud application topologies using cloud elements/services sourced on the fly from multiple providers, based on requirements specified in higher level objectives. Its implementation and use is evaluated in a set of experiments.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125220151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 155
Scan-Sharing for Optimizing RDF Graph Pattern Matching on MapReduce MapReduce上优化RDF图模式匹配的扫描共享
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.14
Hyeongsik Kim, P. Ravindra, Kemafor Anyanwu
Recently, the number and size of RDF data collections has increased rapidly making the issue of scalable processing techniques crucial. The MapReduce model has become a de facto standard for large scale data processing using a cluster of machines in the cloud. Generally, RDF query processing creates join-intensive workloads, resulting in lengthy MapReduce workflows with expensive I/O, data transfer, and sorting costs. However, the MapReduce computation model provides limited static optimization techniques used in relational databases (e.g., indexing and cost-based optimization). Consequently, dynamic optimization techniques for such join-intensive tasks on MapReduce need to be investigated. In some previous efforts, we propose a Nested Triple Group data model and Algebra (NTGA) for efficient graph pattern query processing in the cloud. Here, we extend this work with a scan-sharing technique that is used to optimize the processing of graph patterns with repeated properties. Specifically, our scan-sharing technique eliminates the need for repeated scanning of input relations when properties are used repeatedly in graph patterns. A formal foundation underlying this scan sharing technique is discussed as well as an implementation strategy that has been integrated in the Apache Pig framework is presented. We also present a comprehensive evaluation demonstrating performance benefits of our NTGA plus scan-sharing approach.
最近,RDF数据集合的数量和大小迅速增加,使得可伸缩处理技术的问题变得至关重要。MapReduce模型已经成为使用云中的机器集群进行大规模数据处理的事实上的标准。通常,RDF查询处理会创建连接密集型工作负载,从而导致冗长的MapReduce工作流,并带来昂贵的I/O、数据传输和排序成本。然而,MapReduce计算模型在关系数据库中提供了有限的静态优化技术(例如,索引和基于成本的优化)。因此,需要研究MapReduce上这种连接密集型任务的动态优化技术。在之前的一些工作中,我们提出了一种嵌套三组数据模型和代数(NTGA),用于云中高效的图形模式查询处理。在这里,我们使用扫描共享技术扩展了这项工作,该技术用于优化具有重复属性的图形模式的处理。具体来说,我们的扫描共享技术消除了在图形模式中重复使用属性时重复扫描输入关系的需要。讨论了扫描共享技术的正式基础,并提出了集成在Apache Pig框架中的实现策略。我们还提出了一个全面的评估,展示了我们的NTGA加扫描共享方法的性能优势。
{"title":"Scan-Sharing for Optimizing RDF Graph Pattern Matching on MapReduce","authors":"Hyeongsik Kim, P. Ravindra, Kemafor Anyanwu","doi":"10.1109/CLOUD.2012.14","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.14","url":null,"abstract":"Recently, the number and size of RDF data collections has increased rapidly making the issue of scalable processing techniques crucial. The MapReduce model has become a de facto standard for large scale data processing using a cluster of machines in the cloud. Generally, RDF query processing creates join-intensive workloads, resulting in lengthy MapReduce workflows with expensive I/O, data transfer, and sorting costs. However, the MapReduce computation model provides limited static optimization techniques used in relational databases (e.g., indexing and cost-based optimization). Consequently, dynamic optimization techniques for such join-intensive tasks on MapReduce need to be investigated. In some previous efforts, we propose a Nested Triple Group data model and Algebra (NTGA) for efficient graph pattern query processing in the cloud. Here, we extend this work with a scan-sharing technique that is used to optimize the processing of graph patterns with repeated properties. Specifically, our scan-sharing technique eliminates the need for repeated scanning of input relations when properties are used repeatedly in graph patterns. A formal foundation underlying this scan sharing technique is discussed as well as an implementation strategy that has been integrated in the Apache Pig framework is presented. We also present a comprehensive evaluation demonstrating performance benefits of our NTGA plus scan-sharing approach.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125967764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Minimizing Latency in Serving Requests through Differential Template Caching in a Cloud 通过云中的差异模板缓存来最小化服务请求的延迟
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.17
Deepak Jeswani, Manish Gupta, Pradipta De, Arpit Malani, U. Bellur
In Software-as-a-Service (SaaS) cloud delivery model, a hosting center deploys a Virtual Machine (VM) image template on a server on demand. Image templates are usually maintained in a central repository. With geographically dispersed hosting centers, time to transfer a large, often GigaByte sized, template file from the repository faces high latency due to low Internet bandwidth. An architecture that maintains a template cache, collocated with the hosting centers, can reduce request service latency. Since templates are large in size, caching complete templates is prohibitive in terms of storage space. In order to optimize cache space requirement, as well as, to reduce transfers from the repository, we propose a differential template caching technique, called DiffCache. A difference file or a patch between two templates, that have common components, is small in size. DiffCache computes an optimal selection of templates and patches based on the frequency of requests for specific templates. A template missing in the cache can be generated if any cached template can be patched with a cached patch file, thereby saving the transfer time from the repository at the cost of relatively small patching time. We show that patch based caching coupled with intelligent population of the cache can lead to a 90% improvement in service request latency when compared with caching only template files.
在SaaS (Software-as-a-Service)云交付模式中,托管中心根据需要在服务器上部署虚拟机(Virtual Machine)镜像模板。映像模板通常在中央存储库中维护。对于地理上分散的托管中心,由于Internet带宽较低,从存储库传输大型(通常是gb大小)模板文件的时间面临着高延迟。维护模板缓存的体系结构与托管中心并置,可以减少请求服务延迟。由于模板的大小很大,缓存完整的模板在存储空间方面是令人望而却步的。为了优化缓存空间需求,以及减少存储库的传输,我们提出了一种称为DiffCache的差分模板缓存技术。具有相同组件的两个模板之间的差异文件或补丁的大小很小。DiffCache根据对特定模板的请求频率计算模板和补丁的最佳选择。如果可以使用缓存的补丁文件修补任何缓存的模板,则可以生成缓存中缺失的模板,从而以相对较小的补丁时间为代价,节省从存储库传输的时间。我们表明,与仅缓存模板文件相比,基于补丁的缓存与缓存的智能填充相结合,可以将服务请求延迟提高90%。
{"title":"Minimizing Latency in Serving Requests through Differential Template Caching in a Cloud","authors":"Deepak Jeswani, Manish Gupta, Pradipta De, Arpit Malani, U. Bellur","doi":"10.1109/CLOUD.2012.17","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.17","url":null,"abstract":"In Software-as-a-Service (SaaS) cloud delivery model, a hosting center deploys a Virtual Machine (VM) image template on a server on demand. Image templates are usually maintained in a central repository. With geographically dispersed hosting centers, time to transfer a large, often GigaByte sized, template file from the repository faces high latency due to low Internet bandwidth. An architecture that maintains a template cache, collocated with the hosting centers, can reduce request service latency. Since templates are large in size, caching complete templates is prohibitive in terms of storage space. In order to optimize cache space requirement, as well as, to reduce transfers from the repository, we propose a differential template caching technique, called DiffCache. A difference file or a patch between two templates, that have common components, is small in size. DiffCache computes an optimal selection of templates and patches based on the frequency of requests for specific templates. A template missing in the cache can be generated if any cached template can be patched with a cached patch file, thereby saving the transfer time from the repository at the cost of relatively small patching time. We show that patch based caching coupled with intelligent population of the cache can lead to a 90% improvement in service request latency when compared with caching only template files.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116006518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Impact of Live Migration on Multi-tier Application Performance in Clouds 云环境下热迁移对多层应用性能的影响
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.57
S. Kikuchi, Y. Matsumoto
Live migration technologies can contribute to efficient resource management in a cloud datacenter; however, they will inevitably entail downtime for the virtual machine involved. Even if the downtime is relatively short, its effect can be serious for applications sensitive to response time degradations. Therefore, cloud datacenter providers should control live migration operations to minimize the impact on the performance of applications running on the cloud infrastructure. With this understanding, we studied the impact of live migration on the performance of 2-tier web applications in an experimental setup using XenServer and RUBBoS benchmark. We revealed that the behavior of the transmission control protocol (TCP) can be the primary factor responsible for response time degradation during live migration. On the basis of the experimental results, we constructed functions to estimate the performance impact of live migration on the applications. We also examined a case study to demonstrate how cloud computing datacenters can determine the best live migration strategy to minimize application performance degradation.
实时迁移技术有助于在云数据中心中实现高效的资源管理;然而,它们将不可避免地导致所涉及的虚拟机停机。即使停机时间相对较短,对于对响应时间降低敏感的应用程序,其影响也可能很严重。因此,云数据中心提供商应该控制实时迁移操作,以尽量减少对运行在云基础设施上的应用程序性能的影响。有了这样的理解,我们在使用XenServer和RUBBoS基准的实验设置中研究了实时迁移对两层web应用程序性能的影响。我们发现传输控制协议(TCP)的行为可能是导致实时迁移期间响应时间下降的主要因素。在实验结果的基础上,我们构造了函数来估计实时迁移对应用程序的性能影响。我们还研究了一个案例研究,以演示云计算数据中心如何确定最佳的实时迁移策略,从而最大限度地减少应用程序性能下降。
{"title":"Impact of Live Migration on Multi-tier Application Performance in Clouds","authors":"S. Kikuchi, Y. Matsumoto","doi":"10.1109/CLOUD.2012.57","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.57","url":null,"abstract":"Live migration technologies can contribute to efficient resource management in a cloud datacenter; however, they will inevitably entail downtime for the virtual machine involved. Even if the downtime is relatively short, its effect can be serious for applications sensitive to response time degradations. Therefore, cloud datacenter providers should control live migration operations to minimize the impact on the performance of applications running on the cloud infrastructure. With this understanding, we studied the impact of live migration on the performance of 2-tier web applications in an experimental setup using XenServer and RUBBoS benchmark. We revealed that the behavior of the transmission control protocol (TCP) can be the primary factor responsible for response time degradation during live migration. On the basis of the experimental results, we constructed functions to estimate the performance impact of live migration on the applications. We also examined a case study to demonstrate how cloud computing datacenters can determine the best live migration strategy to minimize application performance degradation.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122491449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Data Centers in the Cloud: A Large Scale Performance Study 云中的数据中心:大规模性能研究
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.87
R. Birke, L. Chen, E. Smirni
With the advancement of virtualization technologies and the benefit of economies of scale, industries are seeking scalable IT solutions, such as data centers hosted either in-house or by a third party. Data center availability, often via a cloud setting, is ubiquitous. Nonetheless, little is known about the in-production performance of data centers, and especially the interaction of workload demands and resource availability. This study fills this gap by conducting a large scale survey of in-production data center servers within a time period that spans two years. We provide in-depth analysis on the time evolution of existing data center demands by providing a holistic characterization of typical data center server workloads, by focusing on their basic resource components, including CPU, memory, and storage systems. We especially focus on seasonality of resource demands and how this is affected by different geographical locations. This survey provides a glimpse on the evolution of data center workloads and provides a basis for an economics analysis that can be used for effective capacity planning of future data centers.
随着虚拟化技术的进步和规模经济的好处,行业正在寻求可扩展的IT解决方案,例如内部托管或由第三方托管的数据中心。数据中心可用性(通常通过云设置)无处不在。尽管如此,对于数据中心的生产性能,特别是工作负载需求和资源可用性之间的相互作用,我们知之甚少。本研究通过在两年的时间内对生产中的数据中心服务器进行大规模调查,填补了这一空白。我们通过提供典型数据中心服务器工作负载的整体特征,重点关注其基本资源组件(包括CPU、内存和存储系统),对现有数据中心需求的时间演变进行了深入分析。我们特别关注资源需求的季节性以及这如何受到不同地理位置的影响。该调查提供了对数据中心工作负载演变的一瞥,并为可用于未来数据中心有效容量规划的经济分析提供了基础。
{"title":"Data Centers in the Cloud: A Large Scale Performance Study","authors":"R. Birke, L. Chen, E. Smirni","doi":"10.1109/CLOUD.2012.87","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.87","url":null,"abstract":"With the advancement of virtualization technologies and the benefit of economies of scale, industries are seeking scalable IT solutions, such as data centers hosted either in-house or by a third party. Data center availability, often via a cloud setting, is ubiquitous. Nonetheless, little is known about the in-production performance of data centers, and especially the interaction of workload demands and resource availability. This study fills this gap by conducting a large scale survey of in-production data center servers within a time period that spans two years. We provide in-depth analysis on the time evolution of existing data center demands by providing a holistic characterization of typical data center server workloads, by focusing on their basic resource components, including CPU, memory, and storage systems. We especially focus on seasonality of resource demands and how this is affected by different geographical locations. This survey provides a glimpse on the evolution of data center workloads and provides a basis for an economics analysis that can be used for effective capacity planning of future data centers.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122994071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 67
Peregrine: An All-Layer-2 Container Computer Network Peregrine:一个全二层容器计算机网络
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.69
T. Chiueh, Cheng-Chun Tu, Yu-Cheng Wang, Pai-Wei Wang, Kai-Wen Li, Yu-Ming Huang
ITRI container computer is a modular computer designed to be a building block for constructing cloud-scale data centers. Rather than using a traditional data center network architecture, which is typically based on a combination of Layer 2 switches and Layer 3 routers, ITRI containercomputer's internal interconnection fabric, called Peregrine, is specially architected to meet the scalability, fast fail-over and multi-tenancy requirements of these data centers. Peregrine is an all-Layer 2 network that is designed to support up to one million Layer 2 end points, provide quick recovery from any single network link/device failure, and incorporate dynamic load-balancing routing to make the best of all physical network links. Finally, the Peregrine architecture is implementable using only off-the-shelf commodity Ethernet switches. This paper describes the design and implementation of a fully operational Peregrine prototype, which is built on a folded Clos physical network topology, and the results and analysis of a performance evaluation study based on measurements taken on this prototype.
工研院容器计算机是一种模块化计算机,设计用于构建云规模的数据中心。传统的数据中心网络架构通常是基于第二层交换机和第三层路由器的组合,而工研院容器计算机的内部互连结构名为Peregrine,专门用于满足这些数据中心的可扩展性、快速故障转移和多租户需求。Peregrine是一个全二层网络,旨在支持多达一百万个二层端点,从任何单个网络链路/设备故障中提供快速恢复,并结合动态负载平衡路由,以充分利用所有物理网络链路。最后,Peregrine架构只能使用现成的商品以太网交换机来实现。本文描述了基于折叠Clos物理网络拓扑的全功能Peregrine原型的设计和实现,以及基于该原型的测量的性能评估研究的结果和分析。
{"title":"Peregrine: An All-Layer-2 Container Computer Network","authors":"T. Chiueh, Cheng-Chun Tu, Yu-Cheng Wang, Pai-Wei Wang, Kai-Wen Li, Yu-Ming Huang","doi":"10.1109/CLOUD.2012.69","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.69","url":null,"abstract":"ITRI container computer is a modular computer designed to be a building block for constructing cloud-scale data centers. Rather than using a traditional data center network architecture, which is typically based on a combination of Layer 2 switches and Layer 3 routers, ITRI containercomputer's internal interconnection fabric, called Peregrine, is specially architected to meet the scalability, fast fail-over and multi-tenancy requirements of these data centers. Peregrine is an all-Layer 2 network that is designed to support up to one million Layer 2 end points, provide quick recovery from any single network link/device failure, and incorporate dynamic load-balancing routing to make the best of all physical network links. Finally, the Peregrine architecture is implementable using only off-the-shelf commodity Ethernet switches. This paper describes the design and implementation of a fully operational Peregrine prototype, which is built on a folded Clos physical network topology, and the results and analysis of a performance evaluation study based on measurements taken on this prototype.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129090481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Lessons Learnt from the Development of GIS Application on Azure Cloud Platform 基于Azure云平台开发GIS应用的经验教训
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.140
Dinesh Agarwal, S. Prasad
Spatial overlay processing is a widely used compute-intensive GIS application that involves aggregation of two or more layers of maps to facilitate intelligent querying on the collocated output data. When large GIS data sets are represented in polygonal (vector) form, spatial analysis runs for extended periods of time, which is undesirable for time-sensitive applications such as emergency response. We have, for the first time, created an open-architecture-based system named Crayons for Azure cloud platform using state-of-the-art techniques. During the course of development of Crayons system, we faced numerous challenges and gained invaluable insights into Azure cloud platform, which are presented in detail in this paper. The challenges range from limitations of cloud storage and computational services to the choices of tools and technologies used for high performance computing (HPC) application design. We report our findings to provide concrete guidelines to an eScience developer for 1) choice of persistent data storage mechanism, 2) data structure representation, 3) communication and synchronization among nodes, 4) building robust failsafe applications, and 5) optimal cost-effective utilization of resources. Our insights into each challenge faced, the solution to overcome it, and the discussion on the lessons learnt from each challenge can be of help to eScience developers starting application development on Azure and possibly other cloud platforms.
空间叠加处理是一种广泛使用的计算密集型GIS应用,它涉及两层或多层地图的聚合,以方便对并置输出数据的智能查询。当大型地理信息系统数据集以多边形(矢量)形式表示时,空间分析运行的时间较长,这对于诸如应急响应等对时间敏感的应用来说是不可取的。我们首次使用最先进的技术,为Azure云平台创建了一个基于开放架构的系统,名为Crayons。在Crayons系统的开发过程中,我们遇到了许多挑战,并获得了对Azure云平台的宝贵见解,本文将详细介绍。挑战的范围从云存储和计算服务的限制到用于高性能计算(HPC)应用程序设计的工具和技术的选择。我们报告了我们的发现,为eScience开发人员提供了具体的指导方针:1)选择持久数据存储机制,2)数据结构表示,3)节点之间的通信和同步,4)构建健壮的故障安全应用程序,以及5)最优的成本效益资源利用。我们对面临的每个挑战的见解、克服这些挑战的解决方案,以及对从每个挑战中吸取的经验教训的讨论,可以帮助eScience开发人员开始在Azure和其他云平台上开发应用程序。
{"title":"Lessons Learnt from the Development of GIS Application on Azure Cloud Platform","authors":"Dinesh Agarwal, S. Prasad","doi":"10.1109/CLOUD.2012.140","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.140","url":null,"abstract":"Spatial overlay processing is a widely used compute-intensive GIS application that involves aggregation of two or more layers of maps to facilitate intelligent querying on the collocated output data. When large GIS data sets are represented in polygonal (vector) form, spatial analysis runs for extended periods of time, which is undesirable for time-sensitive applications such as emergency response. We have, for the first time, created an open-architecture-based system named Crayons for Azure cloud platform using state-of-the-art techniques. During the course of development of Crayons system, we faced numerous challenges and gained invaluable insights into Azure cloud platform, which are presented in detail in this paper. The challenges range from limitations of cloud storage and computational services to the choices of tools and technologies used for high performance computing (HPC) application design. We report our findings to provide concrete guidelines to an eScience developer for 1) choice of persistent data storage mechanism, 2) data structure representation, 3) communication and synchronization among nodes, 4) building robust failsafe applications, and 5) optimal cost-effective utilization of resources. Our insights into each challenge faced, the solution to overcome it, and the discussion on the lessons learnt from each challenge can be of help to eScience developers starting application development on Azure and possibly other cloud platforms.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129649046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
F2Box: Cloudifying F2F Storage Systems with High Availability Correlation F2Box:高可用性相关的F2F存储系统云化
Pub Date : 2012-06-24 DOI: 10.1109/CLOUD.2012.22
Raúl Gracia-Tinedo, Marc S'nchez-Artigas, P. García-López
The increasing popularity of Cloud storage services is leading end-users to store their digital lives (including photos, videos, work documents, etc.) in the Cloud. However, many users are still reluctant to move their data to the Cloud due to the amount of control ceded to Cloud vendors. To let users retain the control over their data, Friend-to-Friend (F2F) storage systems have been presented in the literature as a promising alternative. However, as we show in this paper, pure F2F storage systems present a poor QoS, mainly due to availability correlations, which results in a loss of attractiveness by end users. To overcome this limitation, we propose a hybrid architecture that combines F2F storage systems and the availability of Cloud storage services to let users infer the right balance between user control and quality of service. This architecture, we called it F2BOX, is able to deliver such a balance thanks to the development of a new suite of data transfer scheduling strategies and a new redundancy calculation algorithm. The main feature of this algorithm is that allow users to adjust the amount of redundancy according to the availability patterns exhibited by friends. Our simulation and experimental results (in Amazon S3) demonstrate the high benefits experienced by end users as a result of the "cloudification" of F2F systems.
云存储服务的日益普及正在引导终端用户将他们的数字生活(包括照片、视频、工作文档等)存储在云中。然而,许多用户仍然不愿意将他们的数据迁移到云上,因为云供应商拥有大量的控制权。为了让用户保留对其数据的控制,Friend-to-Friend (F2F)存储系统在文献中被认为是一种很有前途的替代方案。然而,正如我们在本文中所展示的那样,纯粹的F2F存储系统呈现出较差的QoS,主要是由于可用性相关性,这导致最终用户失去吸引力。为了克服这一限制,我们提出了一种混合架构,将F2F存储系统和云存储服务的可用性结合起来,让用户在用户控制和服务质量之间找到适当的平衡。这种架构,我们称之为F2BOX,能够提供这样一种平衡,这要归功于一套新的数据传输调度策略和一个新的冗余计算算法的开发。该算法的主要特点是允许用户根据好友显示的可用性模式来调整冗余的数量。我们的模拟和实验结果(在Amazon S3中)证明了最终用户由于F2F系统的“云化”而体验到的高收益。
{"title":"F2Box: Cloudifying F2F Storage Systems with High Availability Correlation","authors":"Raúl Gracia-Tinedo, Marc S'nchez-Artigas, P. García-López","doi":"10.1109/CLOUD.2012.22","DOIUrl":"https://doi.org/10.1109/CLOUD.2012.22","url":null,"abstract":"The increasing popularity of Cloud storage services is leading end-users to store their digital lives (including photos, videos, work documents, etc.) in the Cloud. However, many users are still reluctant to move their data to the Cloud due to the amount of control ceded to Cloud vendors. To let users retain the control over their data, Friend-to-Friend (F2F) storage systems have been presented in the literature as a promising alternative. However, as we show in this paper, pure F2F storage systems present a poor QoS, mainly due to availability correlations, which results in a loss of attractiveness by end users. To overcome this limitation, we propose a hybrid architecture that combines F2F storage systems and the availability of Cloud storage services to let users infer the right balance between user control and quality of service. This architecture, we called it F2BOX, is able to deliver such a balance thanks to the development of a new suite of data transfer scheduling strategies and a new redundancy calculation algorithm. The main feature of this algorithm is that allow users to adjust the amount of redundancy according to the availability patterns exhibited by friends. Our simulation and experimental results (in Amazon S3) demonstrate the high benefits experienced by end users as a result of the \"cloudification\" of F2F systems.","PeriodicalId":214084,"journal":{"name":"2012 IEEE Fifth International Conference on Cloud Computing","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126806176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
期刊
2012 IEEE Fifth International Conference on Cloud Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1