Pub Date : 2013-12-02DOI: 10.1109/CloudCom.2013.20
Takahiro Hirofuchi, A. Lèbre, L. Pouilloux
Although virtual machine (VM) placement problem has been an active research area over the past decade, the research community is still looking for an open simulation framework that can simulate in an accurate as well as scalable manner VM operations including live migrations. Existing frameworks, however, leverage a naive migration model that considers neither memory update operations nor resource sharing contention, resulting in an underestimate of both the duration of a live migration and the size of migration traffic. In this paper, we propose a simulation framework of virtualized distributed systems with the first class support of live migration operations. We developed a resource share calculation mechanism for VMs and a live migration model implementing the precopy migration algorithm of Qemu/KVM. We extended a widely used simulation toolkit, SimGrid, which allows users to simulate large-scale distributed systems by using user-friendly programming API. Through experiments, we confirmed that our simulation framework correctly reproduced live migration behaviors of the real world under various conditions. Through a first use case, we also confirmed that it is possible to conduct large-scale simulations of complex virtualized workloads upon hundred thousands of VMs upon thousands of physical machines (PMs).
{"title":"Adding a Live Migration Model into SimGrid: One More Step Toward the Simulation of Infrastructure-as-a-Service Concerns","authors":"Takahiro Hirofuchi, A. Lèbre, L. Pouilloux","doi":"10.1109/CloudCom.2013.20","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.20","url":null,"abstract":"Although virtual machine (VM) placement problem has been an active research area over the past decade, the research community is still looking for an open simulation framework that can simulate in an accurate as well as scalable manner VM operations including live migrations. Existing frameworks, however, leverage a naive migration model that considers neither memory update operations nor resource sharing contention, resulting in an underestimate of both the duration of a live migration and the size of migration traffic. In this paper, we propose a simulation framework of virtualized distributed systems with the first class support of live migration operations. We developed a resource share calculation mechanism for VMs and a live migration model implementing the precopy migration algorithm of Qemu/KVM. We extended a widely used simulation toolkit, SimGrid, which allows users to simulate large-scale distributed systems by using user-friendly programming API. Through experiments, we confirmed that our simulation framework correctly reproduced live migration behaviors of the real world under various conditions. Through a first use case, we also confirmed that it is possible to conduct large-scale simulations of complex virtualized workloads upon hundred thousands of VMs upon thousands of physical machines (PMs).","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"234 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131622466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Technological advancements in mobile technology and cloud computing open the door for another paradigm known as Mobile Cloud Computing (MCC). This integration of cloud computing and mobile technology gives numerous facilities to a mobile user, such as the ubiquitous availability of Location Based Services (LBS). The utilization of these LBS services require the knowledge of a user's location, hence threatens the privacy of a user. In this paper we take advantages of ultra-fast processing and reliability of cloud computing and aim to solve the privacy threat issue faced by a mobile user while getting LBS services. We propose a model that utilizes a cloud based server which helps in making of a cloaking region. We highlight that how our model utilizes an untrusted cloud based server and eliminates the use of a trusted anonymizer while providing LBS services to a user securely and anonymously.
{"title":"Towards Achieving Anonymity in LBS: A Cloud Based Untrusted Middleware","authors":"Fizza Abbas, Rasheed Hussain, Junggab Son, Hasoo Eun, Heekuck Oh","doi":"10.1109/CloudCom.2013.143","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.143","url":null,"abstract":"Technological advancements in mobile technology and cloud computing open the door for another paradigm known as Mobile Cloud Computing (MCC). This integration of cloud computing and mobile technology gives numerous facilities to a mobile user, such as the ubiquitous availability of Location Based Services (LBS). The utilization of these LBS services require the knowledge of a user's location, hence threatens the privacy of a user. In this paper we take advantages of ultra-fast processing and reliability of cloud computing and aim to solve the privacy threat issue faced by a mobile user while getting LBS services. We propose a model that utilizes a cloud based server which helps in making of a cloaking region. We highlight that how our model utilizes an untrusted cloud based server and eliminates the use of a trusted anonymizer while providing LBS services to a user securely and anonymously.","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130677178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-12-02DOI: 10.1109/CloudCom.2013.61
Jong-Yul Kim, H. Schulzrinne
Cloud computing is great for scaling applications but the latency in a guest VM can be unpredictable due to resource contention between neighbors. For telephony applications, which are latency-sensitive, we propose a system to monitor telephony server latencies and adapt the server load based on the measured latencies. We implemented the system and evaluated it on an Amazon EC2 test bed. We show indirectly by comparing our server on EC2 and on a local VM, that there may be contention between EC2 VMs in the wild that leads to higher server latency. While there is some overhead due to constant monitoring of the server, our system manages to lower latency by reducing the load to the server.
{"title":"Cloud Support for Latency-Sensitive Telephony Applications","authors":"Jong-Yul Kim, H. Schulzrinne","doi":"10.1109/CloudCom.2013.61","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.61","url":null,"abstract":"Cloud computing is great for scaling applications but the latency in a guest VM can be unpredictable due to resource contention between neighbors. For telephony applications, which are latency-sensitive, we propose a system to monitor telephony server latencies and adapt the server load based on the measured latencies. We implemented the system and evaluated it on an Amazon EC2 test bed. We show indirectly by comparing our server on EC2 and on a local VM, that there may be contention between EC2 VMs in the wild that leads to higher server latency. While there is some overhead due to constant monitoring of the server, our system manages to lower latency by reducing the load to the server.","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126882687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-12-02DOI: 10.1109/CloudCom.2013.126
Mathias Slawik
Contemporary cloud computing solutions incorporate HTTP intermediaries, such as reverse proxies, load balancers, and intrusion prevention systems. These act as TLS server connection ends and access HTTP/TLS plaintext to carry out their functions. This raises many concerns: increased security efforts, the risk of losing confidentiality and integrity, and potentially unauthorized data access. Current HTTP entity-body encryption technologies address these concerns by providing end-to-end security between user agents and origin servers. However, they present disparate deficiencies, e.g., inefficient presentation languages, message-flow vulnerabilities, and the circumvention of HTTP streaming. This paper introduces the Trusted Cloud Transfer Protocol (TCTP), which presents a novel approach to entity-body encryption overcoming these deficiencies. The pivotal idea of TCTP are HTTP application layer encryption channels (HALECs), which integrate TLS functionality into the HTTP application layer. TCTP can be deployed immediately, as it is fully HTTP compliant, and rapidly implemented, as required TLS libraries are widely available. The reliance upon the mature TLS protocol minimizes the risk of introducing new security threats. Furthermore, TLS brings the benefit of relative efficiency, which is demonstrated on the basis of an example TCTP implementation.
{"title":"The Trusted Cloud Transfer Protocol","authors":"Mathias Slawik","doi":"10.1109/CloudCom.2013.126","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.126","url":null,"abstract":"Contemporary cloud computing solutions incorporate HTTP intermediaries, such as reverse proxies, load balancers, and intrusion prevention systems. These act as TLS server connection ends and access HTTP/TLS plaintext to carry out their functions. This raises many concerns: increased security efforts, the risk of losing confidentiality and integrity, and potentially unauthorized data access. Current HTTP entity-body encryption technologies address these concerns by providing end-to-end security between user agents and origin servers. However, they present disparate deficiencies, e.g., inefficient presentation languages, message-flow vulnerabilities, and the circumvention of HTTP streaming. This paper introduces the Trusted Cloud Transfer Protocol (TCTP), which presents a novel approach to entity-body encryption overcoming these deficiencies. The pivotal idea of TCTP are HTTP application layer encryption channels (HALECs), which integrate TLS functionality into the HTTP application layer. TCTP can be deployed immediately, as it is fully HTTP compliant, and rapidly implemented, as required TLS libraries are widely available. The reliance upon the mature TLS protocol minimizes the risk of introducing new security threats. Furthermore, TLS brings the benefit of relative efficiency, which is demonstrated on the basis of an example TCTP implementation.","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126371512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-12-02DOI: 10.1109/CloudCom.2013.158
Ke Huang, Kyrre M. Begnum
This paper demonstrates a bottom-up approach to developing autonomic fault tolerance and disaster recovery on cloud-based deployments. We avoid lock-in to specific recovery features provided by the cloud itself, and instead show that tools used in system administration today can provide the foundation for recovery processes with few additions. The resulting system, Hydra, is capable of detecting failures in instances and can redeploy any instance at a new location without human intervention. The layered design of configuration management tools enables separation of recovery processes and the actual service, making the Hydra applicable to a wide range of scenarios. The implementation was tested and an analysis of the recovery time is provided, demonstrating that the Hydra is capable of completely rebuilding a new site in 15 minutes.
{"title":"The Hydra: A Layered, Redundant Configuration Management Approach for Cloud-Agnostic Disaster Recovery","authors":"Ke Huang, Kyrre M. Begnum","doi":"10.1109/CloudCom.2013.158","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.158","url":null,"abstract":"This paper demonstrates a bottom-up approach to developing autonomic fault tolerance and disaster recovery on cloud-based deployments. We avoid lock-in to specific recovery features provided by the cloud itself, and instead show that tools used in system administration today can provide the foundation for recovery processes with few additions. The resulting system, Hydra, is capable of detecting failures in instances and can redeploy any instance at a new location without human intervention. The layered design of configuration management tools enables separation of recovery processes and the actual service, making the Hydra applicable to a wide range of scenarios. The implementation was tested and an analysis of the recovery time is provided, demonstrating that the Hydra is capable of completely rebuilding a new site in 15 minutes.","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114635671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-12-02DOI: 10.1109/CloudCom.2013.52
Anderson Santana de Oliveira, Jakub Sendor, Alexandr Garaga, Kateline Jenatton
Cloud computing brings a number of compliance risks to organisations because physical perimeters are not clearly delimited. Many regulations relate to the location of the data processing (and storage), including the EU Data protection directive. A major problem for cloud service consumers, acting as data controllers, is how to demonstrate compliance to data transfer constraints. We address the lack of tools to support accountable data localization and transfer across cloud software, platform and infrastructure services, usually run by data processors. In this paper we design a framework for automating the collection of evidence that obligations with respect to personal data handling are being carried out in what concerns personal data transfers. We experiment our approach in the Open Stack open source IaaS implementation, showing how auditors can verify whether data transfers were compliant.
{"title":"Monitoring Personal Data Transfers in the Cloud","authors":"Anderson Santana de Oliveira, Jakub Sendor, Alexandr Garaga, Kateline Jenatton","doi":"10.1109/CloudCom.2013.52","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.52","url":null,"abstract":"Cloud computing brings a number of compliance risks to organisations because physical perimeters are not clearly delimited. Many regulations relate to the location of the data processing (and storage), including the EU Data protection directive. A major problem for cloud service consumers, acting as data controllers, is how to demonstrate compliance to data transfer constraints. We address the lack of tools to support accountable data localization and transfer across cloud software, platform and infrastructure services, usually run by data processors. In this paper we design a framework for automating the collection of evidence that obligations with respect to personal data handling are being carried out in what concerns personal data transfers. We experiment our approach in the Open Stack open source IaaS implementation, showing how auditors can verify whether data transfers were compliant.","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122123010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-12-02DOI: 10.1109/CloudCom.2013.138
Omar Abdul-Rahman, K. Aida
Resource allocation is an active direction of research that is drawing interest within academic and technological circles. Resource allocation imposes numerous challenges. This is especially true for Inter-Clouds, a recent paradigm for horizontal expansion and integration of disparate and heterogeneous cloud platforms. In an attempt to realize an efficient resource management system, this work-in-progress paper proposes and describes a new multi-layered management framework to address the tasks of virtualized resource control, dynamic resource provisioning, life-cycle management and resource exchange within Inter-Cloud environments.
{"title":"Multi-layered Architecture for the Management of Virtualized Application Environments within Inter-cloud Platforms","authors":"Omar Abdul-Rahman, K. Aida","doi":"10.1109/CloudCom.2013.138","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.138","url":null,"abstract":"Resource allocation is an active direction of research that is drawing interest within academic and technological circles. Resource allocation imposes numerous challenges. This is especially true for Inter-Clouds, a recent paradigm for horizontal expansion and integration of disparate and heterogeneous cloud platforms. In an attempt to realize an efficient resource management system, this work-in-progress paper proposes and describes a new multi-layered management framework to address the tasks of virtualized resource control, dynamic resource provisioning, life-cycle management and resource exchange within Inter-Cloud environments.","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122128844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Recently, we have witnessed that cloud providers start to offer heterogeneous computing environments. There have been wide interests in both clusters and cloud of adopting graphics processors (GPUs) as accelerators for various applications. On the other hand, large-scale graph processing is important for many data-intensive applications in the cloud. In this paper, we propose to leverage GPUs to accelerate large-scale graph processing in the cloud. Specifically, we develop an in-memory graph processing engine G2 with three non-trivial GPU-specific optimizations. Firstly, we adopt fine-grained APIs to take advantage of the massive thread parallelism of the GPU. Secondly, G2 embraces a graph partition based approach for load balancing on heterogeneous CPU/GPU architectures. Thirdly, a runtime system is developed to perform transparent memory management on the GPU, and to perform scheduling for an improved throughput of concurrent kernel executions from graph tasks. We have conducted experiments on an Amazon EC2 virtual cluster of eight nodes. Our preliminary results demonstrate that 1) GPU is a viable accelerator for cloud-based graph processing, and 2) the proposed optimizations improve the performance of GPU-based graph processing engine. We further present the lessons learnt and open problems towards large-scale graph processing with GPU accelerations.
{"title":"Towards GPU-Accelerated Large-Scale Graph Processing in the Cloud","authors":"Jianlong Zhong, Bingsheng He","doi":"10.1109/CloudCom.2013.8","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.8","url":null,"abstract":"Recently, we have witnessed that cloud providers start to offer heterogeneous computing environments. There have been wide interests in both clusters and cloud of adopting graphics processors (GPUs) as accelerators for various applications. On the other hand, large-scale graph processing is important for many data-intensive applications in the cloud. In this paper, we propose to leverage GPUs to accelerate large-scale graph processing in the cloud. Specifically, we develop an in-memory graph processing engine G2 with three non-trivial GPU-specific optimizations. Firstly, we adopt fine-grained APIs to take advantage of the massive thread parallelism of the GPU. Secondly, G2 embraces a graph partition based approach for load balancing on heterogeneous CPU/GPU architectures. Thirdly, a runtime system is developed to perform transparent memory management on the GPU, and to perform scheduling for an improved throughput of concurrent kernel executions from graph tasks. We have conducted experiments on an Amazon EC2 virtual cluster of eight nodes. Our preliminary results demonstrate that 1) GPU is a viable accelerator for cloud-based graph processing, and 2) the proposed optimizations improve the performance of GPU-based graph processing engine. We further present the lessons learnt and open problems towards large-scale graph processing with GPU accelerations.","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117283602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-12-02DOI: 10.1109/CloudCom.2013.101
W. Pieters
Cloud architectures are complex socio-technical systems of systems, consisting not only of technological components and their connections, but also of physical premises and employees. When analysing security of such systems and considering countermeasures, the notion of "weakest link" often appears. Humans are then typically said to be the "weakest link" when it comes to security, but no proof is provided for this statement. One reason for this is the fact that there are no unified metrics of security that would apply to physical, digital and social components of complex systems alike. How does one compare the security of a room against the security of a piece of data, and how does social engineering an employee compare to exploiting a server vulnerability? Are we really comparing apples and oranges here, or would it be possible to present a comparative metric that would apply across the different domains? This paper explores the possibility of such a metric for complex systems, and proposes one in terms of the risk induced by an entity in the system. This also provides a foundation for the notion of "weakest link", in terms of the entity (set of entities) with the highest induced risk.
{"title":"Defining \"The Weakest Link\" Comparative Security in Complex Systems of Systems","authors":"W. Pieters","doi":"10.1109/CloudCom.2013.101","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.101","url":null,"abstract":"Cloud architectures are complex socio-technical systems of systems, consisting not only of technological components and their connections, but also of physical premises and employees. When analysing security of such systems and considering countermeasures, the notion of \"weakest link\" often appears. Humans are then typically said to be the \"weakest link\" when it comes to security, but no proof is provided for this statement. One reason for this is the fact that there are no unified metrics of security that would apply to physical, digital and social components of complex systems alike. How does one compare the security of a room against the security of a piece of data, and how does social engineering an employee compare to exploiting a server vulnerability? Are we really comparing apples and oranges here, or would it be possible to present a comparative metric that would apply across the different domains? This paper explores the possibility of such a metric for complex systems, and proposes one in terms of the risk induced by an entity in the system. This also provides a foundation for the notion of \"weakest link\", in terms of the entity (set of entities) with the highest induced risk.","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115053999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2013-12-02DOI: 10.1109/CloudCom.2013.19
L. Assunção, J. Cunha
The workflow paradigm is a well established approach to deal with application complexity by supporting the application development by composition of multiple activities. Furthermore workflows allow encapsulating parts of a problem inside an activity that can be reused in different workflow application scenarios for instance long-running experiments such as the ones involving data streaming. These workflows are characterized by multiple, eventually infinite, iterations processing datasets in multiple activities according to the workflow graph. Some of these activities can invoke Cloud services often unreliably or with limitations on quality of service provoking faults. After a fault the most common approach requires restarting of the entire workflow which can lead to a waste of execution time due to unnecessarily repeating of computations. This paper discuss how the AWARD (Autonomic Workflow Activities Reconfigurable and Dynamic) framework supports recovery from activity faults using dynamic reconfigurations. This is illustrated through an experimental scenario based on a long-running workflow where an activity fails when invoking a Cloud-hosted Web service with a variable level of availability. On detecting this, the AWARD framework allows the dynamic reconfiguration of the corresponding activity to access a new Web service, and avoiding restarting the complete workflow.
工作流范例是一种建立良好的方法,通过支持多个活动组合的应用程序开发来处理应用程序复杂性。此外,工作流允许将问题的部分封装在活动中,这些活动可以在不同的工作流应用程序场景中重用,例如涉及数据流的长时间运行的实验。这些工作流的特点是根据工作流图在多个活动中处理数据集的多次,最终是无限的迭代。其中一些活动可能经常不可靠地调用云服务,或者对服务质量有限制,从而引发故障。在发生故障后,最常见的方法需要重新启动整个工作流,这可能会导致不必要的重复计算而浪费执行时间。本文讨论了AWARD (Autonomic Workflow Activities Reconfigurable and Dynamic)框架如何通过动态重新配置来支持从活动错误中恢复。通过一个基于长时间运行的工作流的实验场景来说明这一点,其中在调用具有可变可用性级别的云托管Web服务时,活动失败。在检测到这一点后,AWARD框架允许动态重新配置相应的活动以访问新的Web服务,并避免重新启动整个工作流。
{"title":"Dynamic Workflow Reconfigurations for Recovering from Faulty Cloud Services","authors":"L. Assunção, J. Cunha","doi":"10.1109/CloudCom.2013.19","DOIUrl":"https://doi.org/10.1109/CloudCom.2013.19","url":null,"abstract":"The workflow paradigm is a well established approach to deal with application complexity by supporting the application development by composition of multiple activities. Furthermore workflows allow encapsulating parts of a problem inside an activity that can be reused in different workflow application scenarios for instance long-running experiments such as the ones involving data streaming. These workflows are characterized by multiple, eventually infinite, iterations processing datasets in multiple activities according to the workflow graph. Some of these activities can invoke Cloud services often unreliably or with limitations on quality of service provoking faults. After a fault the most common approach requires restarting of the entire workflow which can lead to a waste of execution time due to unnecessarily repeating of computations. This paper discuss how the AWARD (Autonomic Workflow Activities Reconfigurable and Dynamic) framework supports recovery from activity faults using dynamic reconfigurations. This is illustrated through an experimental scenario based on a long-running workflow where an activity fails when invoking a Cloud-hosted Web service with a variable level of availability. On detecting this, the AWARD framework allows the dynamic reconfiguration of the corresponding activity to access a new Web service, and avoiding restarting the complete workflow.","PeriodicalId":198053,"journal":{"name":"2013 IEEE 5th International Conference on Cloud Computing Technology and Science","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129720026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}