首页 > 最新文献

Proceedings of the 1st Workshop on High Performance Serverless Computing最新文献

英文 中文
Expanding Cost-Aware Function Execution with Multidimensional Notions of Cost 用多维成本概念拓展成本意识功能执行
Pub Date : 2020-06-25 DOI: 10.1145/3452413.3464790
Matt Baughman, Rohan Kumar, Ian T Foster, K. Chard
Recent advances in networking technology and serverless architectures have enabled automated distribution of compute workloads at the function level. As heterogeneity and physical distribution of computing resources increase, so too does the need to effectively use those resources. This is especially true when leveraging multiple compute resources in the form of local, distributed, and cloud resources. Adding to the complexity of the problem is different notions of "cost" when it comes to using these resources. Tradeoffs exist due to the inherent difference between costs of computation for the end user. For example, deploying a workload on the cloud could be much faster than using local resources but using the cloud incurs a financial cost. Here, the end user is presented with the tradeoff between time and money. We describe preliminary work towards Delts+, a framework that integrates multidimensional cost objectives, cost tradeoffs, and optimization under constraints.
网络技术和无服务器架构的最新进展使计算工作负载能够在功能级别自动分布。随着计算资源的异构性和物理分布的增加,有效使用这些资源的需求也在增加。在利用本地、分布式和云资源形式的多个计算资源时尤其如此。当涉及到使用这些资源时,不同的“成本”概念增加了问题的复杂性。由于最终用户的计算成本之间存在固有的差异,因此存在权衡。例如,在云上部署工作负载可能比使用本地资源快得多,但使用云会产生财务成本。在这里,最终用户需要在时间和金钱之间进行权衡。我们描述了Delts+的初步工作,Delts+是一个集成了多维成本目标、成本权衡和约束下优化的框架。
{"title":"Expanding Cost-Aware Function Execution with Multidimensional Notions of Cost","authors":"Matt Baughman, Rohan Kumar, Ian T Foster, K. Chard","doi":"10.1145/3452413.3464790","DOIUrl":"https://doi.org/10.1145/3452413.3464790","url":null,"abstract":"Recent advances in networking technology and serverless architectures have enabled automated distribution of compute workloads at the function level. As heterogeneity and physical distribution of computing resources increase, so too does the need to effectively use those resources. This is especially true when leveraging multiple compute resources in the form of local, distributed, and cloud resources. Adding to the complexity of the problem is different notions of \"cost\" when it comes to using these resources. Tradeoffs exist due to the inherent difference between costs of computation for the end user. For example, deploying a workload on the cloud could be much faster than using local resources but using the cloud incurs a financial cost. Here, the end user is presented with the tradeoff between time and money. We describe preliminary work towards Delts+, a framework that integrates multidimensional cost objectives, cost tradeoffs, and optimization under constraints.","PeriodicalId":339058,"journal":{"name":"Proceedings of the 1st Workshop on High Performance Serverless Computing","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115064680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Distributed Parallel Analysis Engine for High Energy Physics Using AWS Lambda 基于AWS Lambda的高能物理分布式并行分析引擎
Pub Date : 2020-06-25 DOI: 10.1145/3452413.3464788
Jacek Kusnierz, M. Malawski, V. Padulano, E. T. Saavedra, P. Alonso-Jordá
The High-Energy Physics experiments at CERN produce a high volume of data. It is not possible to analyze big chunks of it within a reasonable time by any single machine. The ROOT framework was recently extended with the distributed computing capabilities for massively parallelized RDataFrame applications. This approach, using the MapReduce pattern underneath, made the heavy computations much more approachable even for the newcomers. This paper explores the possibility of running such analyses on serverless services in public cloud using a purely stateless environment. So far, the distributed approaches used by RDataFrame relied on stateful, fully managed computing frameworks like Apache Spark. Here we show that our newly developed tool is able to use perfectly stateless cloud functions, demonstrating the excellent speedup in parallel stage of processing in our benchmarks.
欧洲核子研究中心的高能物理实验产生了大量的数据。任何一台机器都不可能在合理的时间内分析大量数据。ROOT框架最近被扩展为大规模并行RDataFrame应用程序的分布式计算能力。这种方法使用了底层的MapReduce模式,使得繁重的计算即使对于新手来说也更容易处理。本文探讨了使用纯无状态环境在公共云中无服务器服务上运行此类分析的可能性。到目前为止,RDataFrame使用的分布式方法依赖于有状态的、完全托管的计算框架,比如Apache Spark。在这里,我们展示了我们新开发的工具能够完美地使用无状态云功能,在我们的基准测试中展示了并行处理阶段的出色加速。
{"title":"Distributed Parallel Analysis Engine for High Energy Physics Using AWS Lambda","authors":"Jacek Kusnierz, M. Malawski, V. Padulano, E. T. Saavedra, P. Alonso-Jordá","doi":"10.1145/3452413.3464788","DOIUrl":"https://doi.org/10.1145/3452413.3464788","url":null,"abstract":"The High-Energy Physics experiments at CERN produce a high volume of data. It is not possible to analyze big chunks of it within a reasonable time by any single machine. The ROOT framework was recently extended with the distributed computing capabilities for massively parallelized RDataFrame applications. This approach, using the MapReduce pattern underneath, made the heavy computations much more approachable even for the newcomers. This paper explores the possibility of running such analyses on serverless services in public cloud using a purely stateless environment. So far, the distributed approaches used by RDataFrame relied on stateful, fully managed computing frameworks like Apache Spark. Here we show that our newly developed tool is able to use perfectly stateless cloud functions, demonstrating the excellent speedup in parallel stage of processing in our benchmarks.","PeriodicalId":339058,"journal":{"name":"Proceedings of the 1st Workshop on High Performance Serverless Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130500906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Efficient GPU Sharing for Serverless Workflows 无服务器工作流的高效GPU共享
Pub Date : 2020-06-25 DOI: 10.1145/3452413.3464785
K. Satzke, I. E. Akkus, Ruichuan Chen, Ivica Rimac, M. Stein, Andre Beck, Paarijaat Aditya, M. Vanga, V. Hilt
Serverless computing has emerged as a new cloud computing paradigm, where an application consists of individual functions that can be separately managed and executed. However, the function development environment of all serverless computing frameworks at present is CPU-based. In this paper, we propose to extend the open-sourced KNIX high-performance serverless framework so that it can execute functions on shared GPU cluster resources. We have evaluated the performance impacts on the extended KNIX system by measuring overheads and penalties incurred using different deep learning frameworks.
无服务器计算已经成为一种新的云计算范式,其中应用程序由可以单独管理和执行的单个功能组成。然而,目前所有无服务器计算框架的功能开发环境都是基于cpu的。在本文中,我们提出扩展开源的KNIX高性能无服务器框架,使其能够在共享的GPU集群资源上执行功能。我们通过测量使用不同深度学习框架产生的开销和惩罚来评估对扩展KNIX系统的性能影响。
{"title":"Efficient GPU Sharing for Serverless Workflows","authors":"K. Satzke, I. E. Akkus, Ruichuan Chen, Ivica Rimac, M. Stein, Andre Beck, Paarijaat Aditya, M. Vanga, V. Hilt","doi":"10.1145/3452413.3464785","DOIUrl":"https://doi.org/10.1145/3452413.3464785","url":null,"abstract":"Serverless computing has emerged as a new cloud computing paradigm, where an application consists of individual functions that can be separately managed and executed. However, the function development environment of all serverless computing frameworks at present is CPU-based. In this paper, we propose to extend the open-sourced KNIX high-performance serverless framework so that it can execute functions on shared GPU cluster resources. We have evaluated the performance impacts on the extended KNIX system by measuring overheads and penalties incurred using different deep learning frameworks.","PeriodicalId":339058,"journal":{"name":"Proceedings of the 1st Workshop on High Performance Serverless Computing","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132646548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Apollo: Modular and Distributed Runtime System for Serverless Function Compositions on Cloud, Edge, and IoT Resources Apollo:模块化和分布式运行时系统,用于云、边缘和物联网资源上的无服务器功能组合
Pub Date : 2020-06-25 DOI: 10.1145/3452413.3464793
Fedor Smirnov, Behnaz Pourmohseni, T. Fahringer
This paper provides a first presentation of Apollo, a runtime system for serverless function compositions distributed across the cloud-edge-IoT continuum. Apollo's modular design enables a fine-grained decomposition of the runtime implementation(scheduling, data transmission, etc.) of the application, so that each of the numerous implementation decisions can be optimized separately, fully exploiting the potential for the optimization of the overall performance and costs. Apollo features (a) a flexible model of the application and the available resources and (b) an implementation process based on a large set of independent agents. This flexible structure enables distributing not only the processing, but the implementation process itself across a large number of resources, each running an independent Apollo instance. The ability to flexibly determine the placement of implementation actions opens up new optimization opportunities, while at the same time providing access to greater computing power for optimizing challenging decisions such as task scheduling and the placement and routing of data.
本文首次介绍了Apollo,这是一个分布在云边缘物联网连续体上的无服务器功能组合的运行时系统。Apollo的模块化设计可以对应用程序的运行时实现(调度、数据传输等)进行细粒度的分解,从而可以对众多实现决策中的每一个单独进行优化,充分挖掘整体性能和成本优化的潜力。Apollo的特点是(a)应用程序的灵活模型和可用资源,以及(b)基于大量独立代理的实现过程。这种灵活的结构不仅可以跨大量资源分布处理,还可以跨大量资源分布实现过程本身,每个资源运行一个独立的Apollo实例。灵活地确定实现操作的位置的能力开辟了新的优化机会,同时为优化具有挑战性的决策(如任务调度和数据的放置和路由)提供了更强大的计算能力。
{"title":"Apollo: Modular and Distributed Runtime System for Serverless Function Compositions on Cloud, Edge, and IoT Resources","authors":"Fedor Smirnov, Behnaz Pourmohseni, T. Fahringer","doi":"10.1145/3452413.3464793","DOIUrl":"https://doi.org/10.1145/3452413.3464793","url":null,"abstract":"This paper provides a first presentation of Apollo, a runtime system for serverless function compositions distributed across the cloud-edge-IoT continuum. Apollo's modular design enables a fine-grained decomposition of the runtime implementation(scheduling, data transmission, etc.) of the application, so that each of the numerous implementation decisions can be optimized separately, fully exploiting the potential for the optimization of the overall performance and costs. Apollo features (a) a flexible model of the application and the available resources and (b) an implementation process based on a large set of independent agents. This flexible structure enables distributing not only the processing, but the implementation process itself across a large number of resources, each running an independent Apollo instance. The ability to flexibly determine the placement of implementation actions opens up new optimization opportunities, while at the same time providing access to greater computing power for optimizing challenging decisions such as task scheduling and the placement and routing of data.","PeriodicalId":339058,"journal":{"name":"Proceedings of the 1st Workshop on High Performance Serverless Computing","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133739822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
SLA-aware Workload Scheduling Using Hybrid Cloud Services 使用混合云服务支持sla的工作负载调度
Pub Date : 2020-06-25 DOI: 10.1145/3452413.3464789
Dheeraj Chahal, S. Palepu, Mayank Mishra, Rekha Singhal
Cloud services have an auto-scaling feature for load balancing to meet the performance requirements of an application. Existing auto-scaling techniques are based on upscaling and downscaling cloud resources to distribute the dynamically varying workloads. However, bursty workloads pose many challenges for auto-scaling and sometimes result in Service Level Agreement (SLA) violations. Furthermore, over-provisioning or under-provisioning cloud resources to address dynamically evolving workloads results in performance degradation and cost escalation. In this work, we present a workload characterization based approach for scheduling the bursty workload on a highly scalable serverless architecture in conjunction with a machine learning (ML) platform. We present the use of Amazon Web Services (AWS) ML platform SageMaker and serverless computing platform Lambda for load balancing the inference workload to avoid SLA violations. We evaluate our approach using a recommender system that is based on a deep learning model for inference.
云服务具有用于负载平衡的自动扩展特性,以满足应用程序的性能需求。现有的自动扩展技术是基于升级和缩减云资源来分配动态变化的工作负载。然而,突发工作负载给自动扩展带来了许多挑战,有时会导致违反服务水平协议(SLA)。此外,为解决动态发展的工作负载而过度配置或配置不足的云资源会导致性能下降和成本上升。在这项工作中,我们提出了一种基于工作负载特征的方法,用于与机器学习(ML)平台一起在高度可扩展的无服务器架构上调度突发工作负载。我们介绍了使用Amazon Web Services (AWS) ML平台SageMaker和无服务器计算平台Lambda来负载平衡推理工作负载,以避免违反SLA。我们使用基于深度学习推理模型的推荐系统来评估我们的方法。
{"title":"SLA-aware Workload Scheduling Using Hybrid Cloud Services","authors":"Dheeraj Chahal, S. Palepu, Mayank Mishra, Rekha Singhal","doi":"10.1145/3452413.3464789","DOIUrl":"https://doi.org/10.1145/3452413.3464789","url":null,"abstract":"Cloud services have an auto-scaling feature for load balancing to meet the performance requirements of an application. Existing auto-scaling techniques are based on upscaling and downscaling cloud resources to distribute the dynamically varying workloads. However, bursty workloads pose many challenges for auto-scaling and sometimes result in Service Level Agreement (SLA) violations. Furthermore, over-provisioning or under-provisioning cloud resources to address dynamically evolving workloads results in performance degradation and cost escalation. In this work, we present a workload characterization based approach for scheduling the bursty workload on a highly scalable serverless architecture in conjunction with a machine learning (ML) platform. We present the use of Amazon Web Services (AWS) ML platform SageMaker and serverless computing platform Lambda for load balancing the inference workload to avoid SLA violations. We evaluate our approach using a recommender system that is based on a deep learning model for inference.","PeriodicalId":339058,"journal":{"name":"Proceedings of the 1st Workshop on High Performance Serverless Computing","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130229031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Session details: Session: Full Papers 会议详情:会议:论文全文
K. Chard
{"title":"Session details: Session: Full Papers","authors":"K. Chard","doi":"10.1145/3470764","DOIUrl":"https://doi.org/10.1145/3470764","url":null,"abstract":"","PeriodicalId":339058,"journal":{"name":"Proceedings of the 1st Workshop on High Performance Serverless Computing","volume":"161 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124512793","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Motivating High Performance Serverless Workloads 激发高性能无服务器工作负载
Pub Date : 2020-06-25 DOI: 10.1145/3452413.3464786
H. Nguyen, Zhifei Yang, A. Chien
The historical motivation for serverless comes from internet-of-things, smartphone client server, and the objective of simplifying programming (no provisioning) and scale-down (pay-for-use). These applications are generally low-performance best-effort. However, the serverless model enables flexible software architectures suitable for a wide range of applications that demand high-performance and guaranteed performance. We have studied three such applications - scientific data streaming, virtual/augmented reality, and document annotation. We describe how each can be cast in a serverless software architecture and how the application performance requirements translate into high performance requirements (invocation rate, low and predictable latency) for the underlying serverless system implementation. These applications can require invocations rates as high as millions per second (40 MHz) and latency deadlines below a microsecond (300 ns), and furthermore require performance predictability. All of these capabilities are far in excess of today's commercial serverless offerings and represent interesting research challenges.
无服务器的历史动机来自物联网、智能手机客户端服务器,以及简化编程(无需供应)和缩减规模(按使用付费)的目标。这些应用程序通常是低性能的。然而,无服务器模型支持灵活的软件架构,适用于需要高性能和有保证的性能的广泛应用程序。我们研究了三个这样的应用——科学数据流、虚拟/增强现实和文档注释。我们描述了如何在无服务器软件体系结构中转换它们,以及如何将应用程序性能需求转换为底层无服务器系统实现的高性能需求(调用率、低且可预测的延迟)。这些应用程序可能需要高达每秒百万次(40 MHz)的调用速率和低于1微秒(300 ns)的延迟截止时间,而且还需要性能可预测性。所有这些功能都远远超过了今天的商业无服务器产品,并代表了有趣的研究挑战。
{"title":"Motivating High Performance Serverless Workloads","authors":"H. Nguyen, Zhifei Yang, A. Chien","doi":"10.1145/3452413.3464786","DOIUrl":"https://doi.org/10.1145/3452413.3464786","url":null,"abstract":"The historical motivation for serverless comes from internet-of-things, smartphone client server, and the objective of simplifying programming (no provisioning) and scale-down (pay-for-use). These applications are generally low-performance best-effort. However, the serverless model enables flexible software architectures suitable for a wide range of applications that demand high-performance and guaranteed performance. We have studied three such applications - scientific data streaming, virtual/augmented reality, and document annotation. We describe how each can be cast in a serverless software architecture and how the application performance requirements translate into high performance requirements (invocation rate, low and predictable latency) for the underlying serverless system implementation. These applications can require invocations rates as high as millions per second (40 MHz) and latency deadlines below a microsecond (300 ns), and furthermore require performance predictability. All of these capabilities are far in excess of today's commercial serverless offerings and represent interesting research challenges.","PeriodicalId":339058,"journal":{"name":"Proceedings of the 1st Workshop on High Performance Serverless Computing","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127041491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Towards a Serverless Bioinformatics Cyberinfrastructure Pipeline 迈向无服务器生物信息学网络基础设施管道
Pub Date : 2020-06-25 DOI: 10.1145/3452413.3464787
S. Yao, Muhammad Ali Gulzar, Liqing Zhang, A. Butt
Function-as-a-Service (FaaS) and the serverless computing model offer a powerful abstraction for supporting large-scale applications in the cloud. A major hurdle in this context is that it is non-trivial to transform an application, even an already containerized one, to a FaaS implementation. In this paper, we take the first step towards supporting easier and efficient application transformation to FaaS. We present a systematic scheme to transform applications written in Python into a set of functions that can then be automatically deployed atop platforms such as AWS Lamda. We target a Bioinformatics cyberinfrastructure pipeline, CIWARS, that provides waste-water analysis for the identification of antibiotic-resistant bacteria and viruses such as SARS-CoV-2. Based on our experience with enabling FaaS-based CIWARS, we develop a methodology that would help the conversion of other similar applications to the FaaS model. Our evaluation shows that our approach can correctly transform CIWARS to FaaS, and the new FaaS-based CIWARS incurs only negligible(≤2%) less than 2% overhead for representative workloads.
功能即服务(FaaS)和无服务器计算模型为支持云中的大规模应用程序提供了强大的抽象。这种情况下的一个主要障碍是,将应用程序(即使是已经容器化的应用程序)转换为FaaS实现并非易事。在本文中,我们迈出了第一步,以支持更容易和有效的应用程序到FaaS的转换。我们提出了一个系统的方案,将用Python编写的应用程序转换为一组函数,然后可以自动部署在诸如AWS lambda之类的平台上。我们的目标是生物信息学网络基础设施管道CIWARS,该管道为识别耐抗生素细菌和病毒(如SARS-CoV-2)提供废水分析。根据我们在启用基于FaaS的CIWARS方面的经验,我们开发了一种方法,可以帮助将其他类似的应用程序转换为FaaS模型。我们的评估表明,我们的方法可以正确地将CIWARS转换为FaaS,并且对于代表性工作负载,新的基于FaaS的CIWARS只会产生可忽略不计的(≤2%)不到2%的开销。
{"title":"Towards a Serverless Bioinformatics Cyberinfrastructure Pipeline","authors":"S. Yao, Muhammad Ali Gulzar, Liqing Zhang, A. Butt","doi":"10.1145/3452413.3464787","DOIUrl":"https://doi.org/10.1145/3452413.3464787","url":null,"abstract":"Function-as-a-Service (FaaS) and the serverless computing model offer a powerful abstraction for supporting large-scale applications in the cloud. A major hurdle in this context is that it is non-trivial to transform an application, even an already containerized one, to a FaaS implementation. In this paper, we take the first step towards supporting easier and efficient application transformation to FaaS. We present a systematic scheme to transform applications written in Python into a set of functions that can then be automatically deployed atop platforms such as AWS Lamda. We target a Bioinformatics cyberinfrastructure pipeline, CIWARS, that provides waste-water analysis for the identification of antibiotic-resistant bacteria and viruses such as SARS-CoV-2. Based on our experience with enabling FaaS-based CIWARS, we develop a methodology that would help the conversion of other similar applications to the FaaS model. Our evaluation shows that our approach can correctly transform CIWARS to FaaS, and the new FaaS-based CIWARS incurs only negligible(≤2%) less than 2% overhead for representative workloads.","PeriodicalId":339058,"journal":{"name":"Proceedings of the 1st Workshop on High Performance Serverless Computing","volume":"273 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133505565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Session details: Session: Short Papers 会议详情:会议:简短论文
Zhuozhao Li
{"title":"Session details: Session: Short Papers","authors":"Zhuozhao Li","doi":"10.1145/3470763","DOIUrl":"https://doi.org/10.1145/3470763","url":null,"abstract":"","PeriodicalId":339058,"journal":{"name":"Proceedings of the 1st Workshop on High Performance Serverless Computing","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114633540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 1st Workshop on High Performance Serverless Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1