首页 > 最新文献

Proceedings of the Eighth International Workshop on Serverless Computing最新文献

英文 中文
Transferring transactional business processes to FaaS 将事务性业务流程转移到FaaS
Pub Date : 2022-11-07 DOI: 10.1145/3565382.3565882
Kostas Meladakis, Chrysostomos Zeginis, K. Magoutis, D. Plexousakis
Function-as-a-Service (FaaS) is a modern cloud service model that has gained significant attention from the research and industry communities in recent years for its many benefits such as dynamic scaling, cost efficiency, faster programming, flexibility to microservices and containers technology. However, the building and deployment of serverless applications come with many challenges that need to be tackled, like workflow design complexity and migration of other applications. When transactions between different parties are involved, the workflow becomes knotty and the communication between participants and all properties of transactions have to be properly resolved. Transactions have widely been discussed in Business processes, so same practices might be adopted by serverless workflows. In this work we provide guidelines and mapping mechanisms for transforming transactional Business Process Modeling Notation 2.0 (BPMN2) applications to a serverless platform. We shed light on the current inability of function orchestrators to express workflow definitions, and deal with various architectural dilemmas that stem from the dissimilar nature of stateful BPMN vs. stateless serverless applications. We overcome the unbalanced capabilities between well-established BPMN notations and function orchestration definitions and illustrate how to exploit and combine cloud native services that comes with FaaS to create serverless applications.
功能即服务(FaaS)是一种现代云服务模型,近年来因其诸多优点(如动态扩展、成本效率、更快的编程、对微服务的灵活性和容器技术)而受到了研究和业界的极大关注。然而,无服务器应用程序的构建和部署带来了许多需要解决的挑战,比如工作流设计的复杂性和其他应用程序的迁移。当涉及不同参与方之间的事务时,工作流变得复杂,参与者之间的通信和事务的所有属性都必须得到适当的解决。事务已经在业务流程中得到了广泛的讨论,因此相同的实践可以被无服务器工作流所采用。在这项工作中,我们提供了将事务性业务流程建模符号2.0 (BPMN2)应用程序转换为无服务器平台的指导方针和映射机制。我们阐明了功能编排器目前无法表达工作流定义的问题,并处理了各种架构困境,这些困境源于有状态BPMN与无状态无服务器应用程序的不同性质。我们克服了在完善的BPMN符号和功能编排定义之间不平衡的功能,并说明了如何利用和组合FaaS自带的云原生服务来创建无服务器应用程序。
{"title":"Transferring transactional business processes to FaaS","authors":"Kostas Meladakis, Chrysostomos Zeginis, K. Magoutis, D. Plexousakis","doi":"10.1145/3565382.3565882","DOIUrl":"https://doi.org/10.1145/3565382.3565882","url":null,"abstract":"Function-as-a-Service (FaaS) is a modern cloud service model that has gained significant attention from the research and industry communities in recent years for its many benefits such as dynamic scaling, cost efficiency, faster programming, flexibility to microservices and containers technology. However, the building and deployment of serverless applications come with many challenges that need to be tackled, like workflow design complexity and migration of other applications. When transactions between different parties are involved, the workflow becomes knotty and the communication between participants and all properties of transactions have to be properly resolved. Transactions have widely been discussed in Business processes, so same practices might be adopted by serverless workflows. In this work we provide guidelines and mapping mechanisms for transforming transactional Business Process Modeling Notation 2.0 (BPMN2) applications to a serverless platform. We shed light on the current inability of function orchestrators to express workflow definitions, and deal with various architectural dilemmas that stem from the dissimilar nature of stateful BPMN vs. stateless serverless applications. We overcome the unbalanced capabilities between well-established BPMN notations and function orchestration definitions and illustrate how to exploit and combine cloud native services that comes with FaaS to create serverless applications.","PeriodicalId":132528,"journal":{"name":"Proceedings of the Eighth International Workshop on Serverless Computing","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126495702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Impact of microarchitectural state reuse on serverless functions 微架构状态重用对无服务器功能的影响
Pub Date : 2022-11-07 DOI: 10.1145/3565382.3565879
Truls Asheim, Tanvir Ahmed Khan, Baris Kasicki, Rakesh Kumar
Serverless computing has seen rapid growth in the past few years due to its seamless scalability and zero resource provisioning overhead for developers. In serverless, applications are composed of a set of very short-running functions which are invoked in response to events such as HTTP requests. For better resource utilization, cloud providers interleave the execution of thousands of serverless functions on a single server. Recent work argues that this interleaved execution and short run-times cause the serverless functions to perform poorly on modern processors. This is because interleaved execution thrashes the microarchitectural state of a function, thus forcing its subsequent execution to start from a cold state. Further, due to their short-running nature, serverless functions are unable to amortize the warm-up latency of microarchitectural structures, meaning that most the function execution happen from cold state. In this work, we analyze a function's performance sensitivity to microarchitectural state thrashing induced by interleaved execution. Unlike prior work, our analysis reveals that not all functions experience performance degradation because of microarchitectural state thrashing. The two dominating factors that dictate the impact of thrashing on function performance are function execution time and code footprint. For example, we observe that only the functions with short execution times (< 1 ms) show performance degradation due to thrashing and that this degradation is exacerbated for functions with large code footprints.
无服务器计算由于其无缝可伸缩性和开发人员的零资源配置开销,在过去几年中得到了快速增长。在无服务器中,应用程序由一组运行时间非常短的函数组成,这些函数在响应HTTP请求等事件时被调用。为了更好地利用资源,云提供商在单个服务器上交错执行数千个无服务器功能。最近的研究认为,这种交错执行和较短的运行时间导致无服务器功能在现代处理器上表现不佳。这是因为交错执行会破坏函数的微体系结构状态,从而迫使其后续执行从冷状态开始。此外,由于它们的短时间运行特性,无服务器函数无法分摊微架构结构的预热延迟,这意味着大多数函数执行都是在冷状态下进行的。在这项工作中,我们分析了函数对交错执行引起的微架构状态波动的性能敏感性。与之前的工作不同,我们的分析表明,并非所有函数都会因为微架构状态波动而导致性能下降。决定抖动对函数性能影响的两个主要因素是函数执行时间和代码占用。例如,我们观察到,只有执行时间较短(< 1 ms)的函数才会由于抖动而出现性能下降,而对于代码占用较大的函数,这种下降会加剧。
{"title":"Impact of microarchitectural state reuse on serverless functions","authors":"Truls Asheim, Tanvir Ahmed Khan, Baris Kasicki, Rakesh Kumar","doi":"10.1145/3565382.3565879","DOIUrl":"https://doi.org/10.1145/3565382.3565879","url":null,"abstract":"Serverless computing has seen rapid growth in the past few years due to its seamless scalability and zero resource provisioning overhead for developers. In serverless, applications are composed of a set of very short-running functions which are invoked in response to events such as HTTP requests. For better resource utilization, cloud providers interleave the execution of thousands of serverless functions on a single server. Recent work argues that this interleaved execution and short run-times cause the serverless functions to perform poorly on modern processors. This is because interleaved execution thrashes the microarchitectural state of a function, thus forcing its subsequent execution to start from a cold state. Further, due to their short-running nature, serverless functions are unable to amortize the warm-up latency of microarchitectural structures, meaning that most the function execution happen from cold state. In this work, we analyze a function's performance sensitivity to microarchitectural state thrashing induced by interleaved execution. Unlike prior work, our analysis reveals that not all functions experience performance degradation because of microarchitectural state thrashing. The two dominating factors that dictate the impact of thrashing on function performance are function execution time and code footprint. For example, we observe that only the functions with short execution times (< 1 ms) show performance degradation due to thrashing and that this degradation is exacerbated for functions with large code footprints.","PeriodicalId":132528,"journal":{"name":"Proceedings of the Eighth International Workshop on Serverless Computing","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124876135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
All-you-can-inference: serverless DNN model inference suite All-you-can-inference:无服务器DNN模型推理套件
Pub Date : 2022-11-07 DOI: 10.1145/3565382.3565878
Subin Park, J. Choi, Kyungyong Lee
Serverless computing becomes prevalent and is widely adopted for various applications. Deep learning inference tasks are appropriate to be deployed using a serverless architecture due to the nature of fluctuating task arrival events. When serving a Deep Neural Net (DNN) model in a serverless computing environment, there exist many performance optimization opportunities, including various hardware support, model graph optimization, hardware-agnostic model compilation, memory size and batch size configurations, and many others. Although the serverless computing frees users from cloud resource management overhead, it is still very challenging to find an optimal serverless DNN inference environment among a very large optimization opportunities for the configurations. In this work, we propose All-You-Can-Inference (AYCI), which helps users to find an optimally operating DNN inference in a publicly available serverless computing environment. We have built the proposed system as a service using various fully-managed cloud services and open-sourced the system to help DNN application developers to build an optimal serving environment. The prototype implementation and initial experiment result present the difficulty of finding an optimal DNN inference environment with the varying performance.
无服务器计算越来越流行,并被广泛应用于各种应用程序。由于任务到达事件的波动性,深度学习推理任务适合使用无服务器架构进行部署。在无服务器计算环境中为深度神经网络(DNN)模型提供服务时,存在许多性能优化机会,包括各种硬件支持、模型图优化、与硬件无关的模型编译、内存大小和批处理大小配置等。尽管无服务器计算将用户从云资源管理开销中解放出来,但在配置的大量优化机会中找到最佳的无服务器DNN推理环境仍然非常具有挑战性。在这项工作中,我们提出了All-You-Can-Inference (AYCI),它可以帮助用户在公开可用的无服务器计算环境中找到最佳运行的DNN推理。我们使用各种完全托管的云服务和开源系统来构建提议的系统作为服务,以帮助DNN应用程序开发人员构建最佳的服务环境。原型实现和初步实验结果表明,在性能变化的情况下,很难找到最优的深度神经网络推理环境。
{"title":"All-you-can-inference: serverless DNN model inference suite","authors":"Subin Park, J. Choi, Kyungyong Lee","doi":"10.1145/3565382.3565878","DOIUrl":"https://doi.org/10.1145/3565382.3565878","url":null,"abstract":"Serverless computing becomes prevalent and is widely adopted for various applications. Deep learning inference tasks are appropriate to be deployed using a serverless architecture due to the nature of fluctuating task arrival events. When serving a Deep Neural Net (DNN) model in a serverless computing environment, there exist many performance optimization opportunities, including various hardware support, model graph optimization, hardware-agnostic model compilation, memory size and batch size configurations, and many others. Although the serverless computing frees users from cloud resource management overhead, it is still very challenging to find an optimal serverless DNN inference environment among a very large optimization opportunities for the configurations. In this work, we propose All-You-Can-Inference (AYCI), which helps users to find an optimally operating DNN inference in a publicly available serverless computing environment. We have built the proposed system as a service using various fully-managed cloud services and open-sourced the system to help DNN application developers to build an optimal serving environment. The prototype implementation and initial experiment result present the difficulty of finding an optimal DNN inference environment with the varying performance.","PeriodicalId":132528,"journal":{"name":"Proceedings of the Eighth International Workshop on Serverless Computing","volume":"46 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114307052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Sentinel 哨兵
Pub Date : 2022-11-07 DOI: 10.1145/3565382.3565880
Joe Hattori, S. Kato
Serverless computing is a new computing paradigm that enables application developers to focus on the core service logic of their applications. To protect the host kernel from attacks, serverless service providers need to ensure isolation between application sandboxes while keeping the startup latency and memory usage low. Existing architectures provide the warm start functionality to alleviate the startup latency. However, as warm starts are achieved at the cost of high memory usage, handling all requests as warm starts is virtually impossible. Therefore, mitigating the cold start overhead is key to improving startup latency. In the real world, serverless applications tend to be uncomplicated; they only issue simple system calls and do not modify the underlying filesystem. While existing architectures support full-fledged virtualization, serverless applications support many functionalities and data structures that are unnecessary for simple applications. This paper proposes Sentinel, a serverless architecture to target those simple applications. By stripping the unnecessary functionalities and providing the bare minimum OS virtualization needed to execute the targeted applications, Sentinel achieves drastic improvement compared to existing architectures; up to 10× shorter startup overhead, 8.13× shorter end-to-end execution latency, and 98% lower memory usage.
{"title":"Sentinel","authors":"Joe Hattori, S. Kato","doi":"10.1145/3565382.3565880","DOIUrl":"https://doi.org/10.1145/3565382.3565880","url":null,"abstract":"Serverless computing is a new computing paradigm that enables application developers to focus on the core service logic of their applications. To protect the host kernel from attacks, serverless service providers need to ensure isolation between application sandboxes while keeping the startup latency and memory usage low. Existing architectures provide the warm start functionality to alleviate the startup latency. However, as warm starts are achieved at the cost of high memory usage, handling all requests as warm starts is virtually impossible. Therefore, mitigating the cold start overhead is key to improving startup latency. In the real world, serverless applications tend to be uncomplicated; they only issue simple system calls and do not modify the underlying filesystem. While existing architectures support full-fledged virtualization, serverless applications support many functionalities and data structures that are unnecessary for simple applications. This paper proposes Sentinel, a serverless architecture to target those simple applications. By stripping the unnecessary functionalities and providing the bare minimum OS virtualization needed to execute the targeted applications, Sentinel achieves drastic improvement compared to existing architectures; up to 10× shorter startup overhead, 8.13× shorter end-to-end execution latency, and 98% lower memory usage.","PeriodicalId":132528,"journal":{"name":"Proceedings of the Eighth International Workshop on Serverless Computing","volume":"149 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122399511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Migrating from microservices to serverless: an IoT platform case study 从微服务迁移到无服务器:一个物联网平台案例研究
Pub Date : 2022-10-09 DOI: 10.1145/3565382.3565881
Mohak Chadha, Victor Pacyna, Anshul Jindal, Jiatao Gu, M. Gerndt
Microservice architecture is the common choice for developing cloud applications these days since each individual microservice can be independently modified, replaced, and scaled. As a result, application development and operating cloud infrastructure were bundled together into what is now commonly called DevOps. However, with the increasing popularity of the serverless computing paradigm and its several advantages such as no infrastructure management, a pay-per-use billing policy, and on-demand fine-grained autoscaling, there is a growing interest in utilizing FaaS and server-less CaaS technologies for refactoring microservices-based applications. Towards this, we migrate a complex IoT platform application onto OpenWhisk (OW) and Google Cloud Run (GCR). We comprehensively evaluate the performance of the different deployment strategies, i.e., Google Kubernetes Engine (GKE)-Standard, OW, and GCR for the IoT platform using different load testing scenarios. Results from our experiments show that while GKE standard performs best for most scenarios, GCR is always cheaper wrt costs.
微服务架构是目前开发云应用程序的常用选择,因为每个单独的微服务都可以独立修改、替换和扩展。因此,应用程序开发和操作云基础设施被捆绑在一起,形成了现在通常所说的DevOps。然而,随着无服务器计算范式的日益普及,以及它的一些优点(如无需基础设施管理、按使用付费计费策略和按需细粒度自动伸缩),人们对利用FaaS和无服务器CaaS技术重构基于微服务的应用程序的兴趣越来越大。为此,我们将复杂的物联网平台应用程序迁移到OpenWhisk (OW)和Google Cloud Run (GCR)上。我们全面评估了不同部署策略的性能,即Google Kubernetes引擎(GKE)-Standard, OW和GCR用于物联网平台,使用不同的负载测试场景。我们的实验结果表明,虽然GKE标准在大多数情况下表现最好,但GCR的成本总是更低。
{"title":"Migrating from microservices to serverless: an IoT platform case study","authors":"Mohak Chadha, Victor Pacyna, Anshul Jindal, Jiatao Gu, M. Gerndt","doi":"10.1145/3565382.3565881","DOIUrl":"https://doi.org/10.1145/3565382.3565881","url":null,"abstract":"Microservice architecture is the common choice for developing cloud applications these days since each individual microservice can be independently modified, replaced, and scaled. As a result, application development and operating cloud infrastructure were bundled together into what is now commonly called DevOps. However, with the increasing popularity of the serverless computing paradigm and its several advantages such as no infrastructure management, a pay-per-use billing policy, and on-demand fine-grained autoscaling, there is a growing interest in utilizing FaaS and server-less CaaS technologies for refactoring microservices-based applications. Towards this, we migrate a complex IoT platform application onto OpenWhisk (OW) and Google Cloud Run (GCR). We comprehensively evaluate the performance of the different deployment strategies, i.e., Google Kubernetes Engine (GKE)-Standard, OW, and GCR for the IoT platform using different load testing scenarios. Results from our experiments show that while GKE standard performs best for most scenarios, GCR is always cheaper wrt costs.","PeriodicalId":132528,"journal":{"name":"Proceedings of the Eighth International Workshop on Serverless Computing","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116994027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Proceedings of the Eighth International Workshop on Serverless Computing 第八届无服务器计算国际研讨会论文集
{"title":"Proceedings of the Eighth International Workshop on Serverless Computing","authors":"","doi":"10.1145/3565382","DOIUrl":"https://doi.org/10.1145/3565382","url":null,"abstract":"","PeriodicalId":132528,"journal":{"name":"Proceedings of the Eighth International Workshop on Serverless Computing","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120977252","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the Eighth International Workshop on Serverless Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1