首页 > 最新文献

2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)最新文献

英文 中文
Self-Aware Workload Forecasting in Data Center Power Prediction 数据中心功率预测中的自感知工作负荷预测
Ying-Feng Hsu, Kazuhiro Matsuda, Morito Matsuoka
The number and scale of data centers are rapidly increasing, due to the growing demand for cloud computing services. Cloud computing infrastructure relies on a massive amount of information and communication technology (ICT) equipment, which consume an enormous amount of power. Power saving and energy optimization have therefore become essential goals for data centers. An enhanced data center energy management system (DEMS) provides a solution for data center power consumption based on its coordinative control of ICT equipment. An efficient power prediction model is essential for such a DEMS because it facilitates the proactive control of ICT equipment and reduces the total power consumption. In this paper, we propose a novel self-aware workload forecasting (SAWF) framework for total power consumption prediction in data centers. It includes three major components. First, there is a feature selection module, which evaluates the importance of variables from all ICT equipment in a data center and dynamically selects the most relevant variables for data input. Second, we propose an accurate and efficient neural network model to forecast future total power consumption. Third, we provide an online error monitoring and model updating module that continuously monitors prediction errors and updates the model when necessary.
由于对云计算服务的需求不断增长,数据中心的数量和规模正在迅速增加。云计算基础设施依赖于大量的信息和通信技术(ICT)设备,这些设备消耗了大量的电力。因此,节能和能源优化已成为数据中心的基本目标。增强型数据中心能源管理系统(dem)通过对ICT设备的协同控制,为数据中心的能耗问题提供解决方案。一个有效的功率预测模型对于这样的dem至关重要,因为它有助于对ICT设备的主动控制并降低总功耗。在本文中,我们提出了一种新的自我感知工作负载预测(SAWF)框架,用于数据中心总功耗预测。它包括三个主要组成部分。首先,有一个特征选择模块,它评估数据中心中所有ICT设备变量的重要性,并动态选择最相关的变量进行数据输入。其次,我们提出了一个准确、高效的神经网络模型来预测未来的总功耗。第三,我们提供了一个在线误差监测和模型更新模块,持续监测预测误差并在必要时更新模型。
{"title":"Self-Aware Workload Forecasting in Data Center Power Prediction","authors":"Ying-Feng Hsu, Kazuhiro Matsuda, Morito Matsuoka","doi":"10.1109/CCGRID.2018.00047","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00047","url":null,"abstract":"The number and scale of data centers are rapidly increasing, due to the growing demand for cloud computing services. Cloud computing infrastructure relies on a massive amount of information and communication technology (ICT) equipment, which consume an enormous amount of power. Power saving and energy optimization have therefore become essential goals for data centers. An enhanced data center energy management system (DEMS) provides a solution for data center power consumption based on its coordinative control of ICT equipment. An efficient power prediction model is essential for such a DEMS because it facilitates the proactive control of ICT equipment and reduces the total power consumption. In this paper, we propose a novel self-aware workload forecasting (SAWF) framework for total power consumption prediction in data centers. It includes three major components. First, there is a feature selection module, which evaluates the importance of variables from all ICT equipment in a data center and dynamically selects the most relevant variables for data input. Second, we propose an accurate and efficient neural network model to forecast future total power consumption. Third, we provide an online error monitoring and model updating module that continuously monitors prediction errors and updates the model when necessary.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"94 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132320659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
IoT Edge Device Based Key Frame Extraction for Face in Video Recognition 基于IoT边缘设备的视频识别人脸关键帧提取
Xuan Qi, Chen Liu, S. Schuckers
Following the development of computing and communication technologies, the idea of Internet of Things (IoT) has been realized not only at research level but also at application level. Among various IoT-related application fields, biometrics applications, especially face recognition, are widely applied in video-based surveillance, access control, law enforcement and many other scenarios. In this paper, we introduce a Face in Video Recognition (FivR) framework which performs real-time key-frame extraction on IoT edge devices, then conduct face recognition using the extracted key-frames on the Cloud back-end. With our key-frame extraction engine, we are able to reduce the data volume hence dramatically relief the processing pressure of the cloud back-end. Our experimental results show with IoT edge device acceleration, it is possible to implement face in video recognition application without introducing the middle-ware or cloud-let layer, while still achieving real-time processing speed.
随着计算和通信技术的发展,物联网(IoT)的概念不仅在研究层面得到了实现,而且在应用层面得到了实现。在物联网相关的诸多应用领域中,生物识别应用,尤其是人脸识别,被广泛应用于基于视频的监控、门禁、执法等诸多场景。在本文中,我们介绍了一个人脸视频识别(FivR)框架,该框架在物联网边缘设备上进行实时关键帧提取,然后在云后端使用提取的关键帧进行人脸识别。通过我们的关键帧提取引擎,我们能够减少数据量,从而大大减轻云后端的处理压力。我们的实验结果表明,通过物联网边缘设备加速,可以在不引入中间件或云let层的情况下实现视频识别应用中的人脸,同时仍然可以实现实时处理速度。
{"title":"IoT Edge Device Based Key Frame Extraction for Face in Video Recognition","authors":"Xuan Qi, Chen Liu, S. Schuckers","doi":"10.1109/CCGRID.2018.00087","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00087","url":null,"abstract":"Following the development of computing and communication technologies, the idea of Internet of Things (IoT) has been realized not only at research level but also at application level. Among various IoT-related application fields, biometrics applications, especially face recognition, are widely applied in video-based surveillance, access control, law enforcement and many other scenarios. In this paper, we introduce a Face in Video Recognition (FivR) framework which performs real-time key-frame extraction on IoT edge devices, then conduct face recognition using the extracted key-frames on the Cloud back-end. With our key-frame extraction engine, we are able to reduce the data volume hence dramatically relief the processing pressure of the cloud back-end. Our experimental results show with IoT edge device acceleration, it is possible to implement face in video recognition application without introducing the middle-ware or cloud-let layer, while still achieving real-time processing speed.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131366063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Multi-Level Load Balancing with an Integrated Runtime Approach 集成运行时方法的多级负载平衡
Seonmyeong Bak, Harshitha Menon, Sam White, M. Diener, L. Kalé
The recent trend of increasing numbers of cores per chip has resulted in vast amounts of on-node parallelism. These high core counts result in hardware variability that introduces imbalance. Applications are also becoming more complex, re-sulting in dynamic load imbalance. Load imbalance of any kind can result in loss of performance and system utilization. We address the challenge of handling both transient and persistent load imbalances while maintaining locality with low overhead. In this paper, we propose an integrated runtime system that combines the Charm++ distributed programming model with concurrent tasks to mitigate load imbalances within and across shared memory address spaces. It utilizes a periodic assignment of work to cores based on load measurement, in combination with user created tasks to handle load imbalance. We integrate OpenMP with Charm++ to enable creation of potential tasks via OpenMP's parallel loop construct. This is also available to MPI applications through the Adaptive MPI implementation. We demonstrate the benefits of our work on three applications. We show improvements of Lassen by 29.6% on Cori and 46.5% on Theta. We also demonstrate the benefits on a Charm++ application, ChaNGa by 25.7% on Theta, as well as an MPI proxy application, Kripke, using Adaptive MPI.
最近每个芯片的核心数量不断增加的趋势导致了大量的节点上并行性。这些高核数导致硬件可变性,从而导致不平衡。应用程序也变得越来越复杂,导致动态负载不平衡。任何类型的负载不平衡都可能导致性能和系统利用率的损失。我们解决了处理瞬时和持久负载不平衡的挑战,同时保持低开销的局部性。在本文中,我们提出了一个集成的运行时系统,该系统将Charm++分布式编程模型与并发任务相结合,以减轻共享内存地址空间内和跨地址空间的负载不平衡。它利用基于负载测量的周期性工作分配给核心,并结合用户创建的任务来处理负载不平衡。我们将OpenMP与Charm++集成在一起,通过OpenMP的并行循环结构创建潜在任务。这也可以通过自适应MPI实现提供给MPI应用程序。我们将在三个应用程序上演示我们的工作带来的好处。我们显示Lassen在Cori和Theta上分别提高了29.6%和46.5%。我们还演示了在Theta上提高25.7%的Charm++应用程序ChaNGa以及使用自适应MPI的MPI代理应用程序Kripke的好处。
{"title":"Multi-Level Load Balancing with an Integrated Runtime Approach","authors":"Seonmyeong Bak, Harshitha Menon, Sam White, M. Diener, L. Kalé","doi":"10.1109/CCGRID.2018.00018","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00018","url":null,"abstract":"The recent trend of increasing numbers of cores per chip has resulted in vast amounts of on-node parallelism. These high core counts result in hardware variability that introduces imbalance. Applications are also becoming more complex, re-sulting in dynamic load imbalance. Load imbalance of any kind can result in loss of performance and system utilization. We address the challenge of handling both transient and persistent load imbalances while maintaining locality with low overhead. In this paper, we propose an integrated runtime system that combines the Charm++ distributed programming model with concurrent tasks to mitigate load imbalances within and across shared memory address spaces. It utilizes a periodic assignment of work to cores based on load measurement, in combination with user created tasks to handle load imbalance. We integrate OpenMP with Charm++ to enable creation of potential tasks via OpenMP's parallel loop construct. This is also available to MPI applications through the Adaptive MPI implementation. We demonstrate the benefits of our work on three applications. We show improvements of Lassen by 29.6% on Cori and 46.5% on Theta. We also demonstrate the benefits on a Charm++ application, ChaNGa by 25.7% on Theta, as well as an MPI proxy application, Kripke, using Adaptive MPI.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124014066","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
A Conceptual Framework for the Use of Graph Representation Within High Energy Physics Analysis 在高能物理分析中使用图表示的概念框架
Danielle Turvill, L. Barnby, A. Anjum
A new method is presented for improvement of the particle identification analysis process in a way which combines both the measured features, from detectors, and physics parameters. It is proposed that a graph representation can effectively express data in a format allowing for simpler interpretation and exploitation of all data available for analysis purposes. Nodes will represent entities and edges will represent the relation between them. Not only are graphs able to provide this useful structure and formal representation of knowledge but they can also be managed efficiently. Overall, this graphical representation will allow for the study of relationships between tracks, enable better pattern recognition and, as a result, improve the classification of particles.
提出了一种结合探测器测量特征和物理参数的粒子识别分析方法。提出图形表示可以有效地以一种格式表示数据,允许更简单的解释和利用所有可用于分析目的的数据。节点表示实体,边表示实体之间的关系。图不仅能够提供这种有用的结构和知识的正式表示,而且还可以有效地管理它们。总的来说,这种图形表示将允许研究轨迹之间的关系,实现更好的模式识别,从而改进粒子的分类。
{"title":"A Conceptual Framework for the Use of Graph Representation Within High Energy Physics Analysis","authors":"Danielle Turvill, L. Barnby, A. Anjum","doi":"10.1109/CCGRID.2018.00063","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00063","url":null,"abstract":"A new method is presented for improvement of the particle identification analysis process in a way which combines both the measured features, from detectors, and physics parameters. It is proposed that a graph representation can effectively express data in a format allowing for simpler interpretation and exploitation of all data available for analysis purposes. Nodes will represent entities and edges will represent the relation between them. Not only are graphs able to provide this useful structure and formal representation of knowledge but they can also be managed efficiently. Overall, this graphical representation will allow for the study of relationships between tracks, enable better pattern recognition and, as a result, improve the classification of particles.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129971488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Scalable Cloud-Edge Computing Framework for Supporting Device-Adaptive Big Media Provisioning 支持设备自适应大媒体供应的可扩展云边缘计算框架
A. Galletta, A. Cuzzocrea, A. Celesti, M. Fazio, M. Villari
Nowadays, we are observing an explosion on recording and transmitting of videos from multiple sources such as Social Media (Periscope, Facebook, Youtube etc.) and owner of trains/coaches (Trenitalia-Frecce-Italy, TGV-France, Ryanair-Bus-Travels, etc.). In this paper, we investigate how to support the provisioning of videos to heterogeneous end user devices in different contexts, adapting the content to the specific requirements of the used end devices. In particular, we present a new Cloud-Edge Service for vIdeO delivery (CESIO) architecture, that exploits Cloud and Edge virtual resources to improve the delivery video contents at different quality resolutions. The paper describes architecture components and their behaviour in the system. Moreover, a possible application scenario is discussed to well explain how the proposed solution works.
如今,我们正在观察到从社交媒体(Periscope, Facebook, Youtube等)和火车/客车所有者(Trenitalia-Frecce-Italy, TGV-France, Ryanair-Bus-Travels等)等多种来源录制和传输视频的爆炸式增长。在本文中,我们研究了如何支持在不同环境下向异构终端用户设备提供视频,使内容适应所使用的终端设备的特定需求。特别地,我们提出了一种新的云边缘服务视频交付(CESIO)架构,它利用云和边缘虚拟资源来改进不同质量分辨率下的视频内容交付。本文描述了体系结构组件及其在系统中的行为。此外,还讨论了一个可能的应用场景,以很好地解释所建议的解决方案是如何工作的。
{"title":"A Scalable Cloud-Edge Computing Framework for Supporting Device-Adaptive Big Media Provisioning","authors":"A. Galletta, A. Cuzzocrea, A. Celesti, M. Fazio, M. Villari","doi":"10.1109/CCGRID.2018.00099","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00099","url":null,"abstract":"Nowadays, we are observing an explosion on recording and transmitting of videos from multiple sources such as Social Media (Periscope, Facebook, Youtube etc.) and owner of trains/coaches (Trenitalia-Frecce-Italy, TGV-France, Ryanair-Bus-Travels, etc.). In this paper, we investigate how to support the provisioning of videos to heterogeneous end user devices in different contexts, adapting the content to the specific requirements of the used end devices. In particular, we present a new Cloud-Edge Service for vIdeO delivery (CESIO) architecture, that exploits Cloud and Edge virtual resources to improve the delivery video contents at different quality resolutions. The paper describes architecture components and their behaviour in the system. Moreover, a possible application scenario is discussed to well explain how the proposed solution works.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126174216","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
An Elasticity Study of Distributed Graph Processing 分布式图处理的弹性研究
Sietse Au, Alexandru Uta, A. Ilyushkin, A. Iosup
Graphs are a natural fit for modeling concepts used in solving diverse problems in science, commerce, engineering, and governance. Responding to the variety of graph data and algorithms, many parallel and distributed graph processing systems exist. However, until now these platforms use a static model of deployment: they only run on a pre-defined set of machines. This raises many conceptual and pragmatic issues, including misfit with the highly dynamic nature of graph processing, and could lead to resource waste and high operational costs. In contrast, in this work we explore a dynamic model of deployment. We first characterize workload dynamicity, beyond mere active-vertex variability. Then, to conduct an in-depth elasticity study of distributed graph processing, we build a prototype, JoyGraph, which is the first such system that implements complex, policy-based, and fine-grained elasticity. Using the state-of-the-art LDBC Graphalytics benchmark and the SPEC Cloud Group's elasticity metrics, we show the benefits of elasticity in graph processing: (i) improved resource utilization, (ii) reduced operational costs, and (iii) aligned operation-workload dynamicity. Furthermore, we explore the cost of elasticity in graph processing. We identify a key drawback: although elasticity does not degrade application throughput, graph-processing workloads are sensitive to data movement while leasing or releasing resources.
图非常适合用于解决科学、商业、工程和治理中的各种问题的概念建模。为了应对图形数据和算法的多样性,存在许多并行和分布式图形处理系统。然而,到目前为止,这些平台使用的是静态部署模型:它们只运行在一组预定义的机器上。这引发了许多概念和实际问题,包括与图形处理的高度动态特性不匹配,并可能导致资源浪费和高运营成本。相反,在这项工作中,我们探索了一个动态的部署模型。我们首先描述工作负载的动态性,而不仅仅是活动顶点的可变性。然后,为了深入研究分布式图形处理的弹性,我们构建了一个原型JoyGraph,这是第一个实现复杂的、基于策略的、细粒度弹性的系统。使用最先进的LDBC graphhalytics基准测试和SPEC Cloud Group的弹性指标,我们展示了弹性在图形处理中的好处:(i)提高资源利用率,(ii)降低运营成本,以及(iii)调整操作工作负载动态。此外,我们还探讨了图处理中的弹性代价。我们发现了一个关键的缺点:尽管弹性不会降低应用程序吞吐量,但图形处理工作负载在租用或释放资源时对数据移动很敏感。
{"title":"An Elasticity Study of Distributed Graph Processing","authors":"Sietse Au, Alexandru Uta, A. Ilyushkin, A. Iosup","doi":"10.1109/CCGRID.2018.00062","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00062","url":null,"abstract":"Graphs are a natural fit for modeling concepts used in solving diverse problems in science, commerce, engineering, and governance. Responding to the variety of graph data and algorithms, many parallel and distributed graph processing systems exist. However, until now these platforms use a static model of deployment: they only run on a pre-defined set of machines. This raises many conceptual and pragmatic issues, including misfit with the highly dynamic nature of graph processing, and could lead to resource waste and high operational costs. In contrast, in this work we explore a dynamic model of deployment. We first characterize workload dynamicity, beyond mere active-vertex variability. Then, to conduct an in-depth elasticity study of distributed graph processing, we build a prototype, JoyGraph, which is the first such system that implements complex, policy-based, and fine-grained elasticity. Using the state-of-the-art LDBC Graphalytics benchmark and the SPEC Cloud Group's elasticity metrics, we show the benefits of elasticity in graph processing: (i) improved resource utilization, (ii) reduced operational costs, and (iii) aligned operation-workload dynamicity. Furthermore, we explore the cost of elasticity in graph processing. We identify a key drawback: although elasticity does not degrade application throughput, graph-processing workloads are sensitive to data movement while leasing or releasing resources.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"242 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116046635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Hard Real-time Scheduler for Spark on YARN 基于YARN的Spark硬实时调度器
Guolu Wang, Jungang Xu, Renfeng Liu, Shanshan Huang
Apache Spark is a fast and general engine for large-scale data processing using distributed memory. It provides different deploy modes to meet the needs of different users and Spark on YARN is the most popular deploy mode. Different deploy modes have different scheduling mechanisms. Spark on YARN has three different schedulers, including FIFO Scheduler, Fair Scheduler, and Capacity Scheduler. However, these three schedulers cannot fit hard real-time application scenarios. With the application of Apache Spark more widely, the needs of hard real-time scheduling will increase quickly. In this paper, we proposed a novel hard real-time scheduling algorithm called DVDA (Deadline and Value Density-Aware) in order to meet the requirements of hard real-time scheduling. Compared with traditional EDF (Earliest Deadline First) algorithm which only considers the deadline, the DVDA algorithm considers both the deadline and value density of the application. Furthermore, we implement a DVDA Scheduler for Spark on YARN based on the DVDA algorithm. Finally, the experiments are conducted to verify the effectiveness of the algorithm. Experimental results show that the proposed algorithm can increase the application completed rate by 18% and 6%, Value Income by 78% and 32% compared with default Capacity scheduler and EDF-Capacity scheduler respectively.
Apache Spark是一个使用分布式内存进行大规模数据处理的快速通用引擎。它提供了不同的部署模式,以满足不同用户的需求,其中Spark on YARN是最流行的部署模式。不同的部署模式有不同的调度机制。Spark on YARN有三种不同的调度器,包括FIFO调度器、公平调度器和容量调度器。但是,这三个调度器不适合硬实时应用程序场景。随着Apache Spark的应用越来越广泛,对硬实时调度的需求也将迅速增加。为了满足硬实时调度的要求,本文提出了一种新的硬实时调度算法DVDA (Deadline and Value Density-Aware)。与传统的EDF(最早截止日期优先)算法只考虑截止日期相比,DVDA算法同时考虑了应用程序的截止日期和值密度。此外,我们还基于DVDA算法在YARN上实现了Spark的DVDA调度程序。最后通过实验验证了算法的有效性。实验结果表明,与默认Capacity scheduler和EDF-Capacity scheduler相比,该算法的应用完成率分别提高18%和6%,Value Income分别提高78%和32%。
{"title":"A Hard Real-time Scheduler for Spark on YARN","authors":"Guolu Wang, Jungang Xu, Renfeng Liu, Shanshan Huang","doi":"10.1109/CCGRID.2018.00096","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00096","url":null,"abstract":"Apache Spark is a fast and general engine for large-scale data processing using distributed memory. It provides different deploy modes to meet the needs of different users and Spark on YARN is the most popular deploy mode. Different deploy modes have different scheduling mechanisms. Spark on YARN has three different schedulers, including FIFO Scheduler, Fair Scheduler, and Capacity Scheduler. However, these three schedulers cannot fit hard real-time application scenarios. With the application of Apache Spark more widely, the needs of hard real-time scheduling will increase quickly. In this paper, we proposed a novel hard real-time scheduling algorithm called DVDA (Deadline and Value Density-Aware) in order to meet the requirements of hard real-time scheduling. Compared with traditional EDF (Earliest Deadline First) algorithm which only considers the deadline, the DVDA algorithm considers both the deadline and value density of the application. Furthermore, we implement a DVDA Scheduler for Spark on YARN based on the DVDA algorithm. Finally, the experiments are conducted to verify the effectiveness of the algorithm. Experimental results show that the proposed algorithm can increase the application completed rate by 18% and 6%, Value Income by 78% and 32% compared with default Capacity scheduler and EDF-Capacity scheduler respectively.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"131 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127023801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Davram: Distributed Virtual Memory in User Space Davram:用户空间中的分布式虚拟内存
Linhua Jiang, Ke Wang, Dongfang Zhao
One of the most challenging problems in modern distributed big data systems lies in their memory management: these systems preallocate a fixed amount of memory before applications start. In the best case where more memory can be acquired, users have to reconfigure the deployment and re-compute many intermediate results. If no more memory is available, users are then forced to manually partition the job into smaller tasks, incurring both development and performance overhead. This paper presents a user-level utility for managing the memory in distributed systems—the Distributed and Autonomous Virtual RAM (Davram). Davram enables to efficiently swap data between memory and disk in a distributed system without users' intervention or applications' awareness.
现代分布式大数据系统中最具挑战性的问题之一在于它们的内存管理:这些系统在应用程序启动之前预先分配了固定数量的内存。在可以获得更多内存的最佳情况下,用户必须重新配置部署并重新计算许多中间结果。如果没有更多的可用内存,那么用户将被迫手动将作业划分为更小的任务,从而导致开发和性能开销。本文提出了一种用于管理分布式系统中内存的用户级实用程序——分布式自治虚拟内存(Davram)。Davram能够在分布式系统中有效地在内存和磁盘之间交换数据,而无需用户干预或应用程序的意识。
{"title":"Davram: Distributed Virtual Memory in User Space","authors":"Linhua Jiang, Ke Wang, Dongfang Zhao","doi":"10.1109/CCGRID.2018.00050","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00050","url":null,"abstract":"One of the most challenging problems in modern distributed big data systems lies in their memory management: these systems preallocate a fixed amount of memory before applications start. In the best case where more memory can be acquired, users have to reconfigure the deployment and re-compute many intermediate results. If no more memory is available, users are then forced to manually partition the job into smaller tasks, incurring both development and performance overhead. This paper presents a user-level utility for managing the memory in distributed systems—the Distributed and Autonomous Virtual RAM (Davram). Davram enables to efficiently swap data between memory and disk in a distributed system without users' intervention or applications' awareness.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132535412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Provider-Agnostic Approach to Multi-cloud Orchestration Using a Constraint Language 使用约束语言实现多云编排的提供者不可知方法
Dani Baur, Daniel Seybold, F. Griesinger, Hynek Masata, Jörg Domaschka
Cloud computing and its computing as an utility paradigm provides on-demand resources allowing the seamless adaptation of applications to fluctuating demands. While the Cloud's ongoing commercialisation has lead to a vast provider landscape, vendor lock-in is still a major hindrance. Recent outages demonstrate that relying exclusively on one provider is not sufficient. While existing cloud orchestration tools promise to solve the problems by supporting deployments across multiple cloud providers, they typically rely on provider dependent models forcing prior knowledge of offers and obstructing flexibility in case of errors. We propose a cloud provider-agnostic application and resource description using a constraint language. It allows users to express resource requirements of an application without prior knowledge of existing offers. Additionally, we propose a discovery service automatically collecting available offers. We combine this with a matchmaking algorithm representing the discovery model and the user-given constraints in a constraint satisfaction problem (CSP) that is then solved. Finally, we manipulate this discovery model during runtime to react on errors. Our evaluation shows that using a constraint-based language is a feasible approach to the provider selection problem, and that it helps to overcome vendor lock-in.
云计算及其作为实用程序范例的计算提供了按需资源,允许应用程序无缝地适应波动的需求。虽然云的持续商业化已经带来了一个巨大的供应商景观,但供应商锁定仍然是一个主要障碍。最近的中断表明,完全依赖一个提供者是不够的。虽然现有的云编排工具承诺通过支持跨多个云提供商的部署来解决问题,但它们通常依赖于依赖于提供商的模型,强制事先了解提供的服务,并且在出现错误时阻碍灵活性。我们提出了一个云提供商无关的应用程序和使用约束语言的资源描述。它允许用户表达应用程序的资源需求,而无需事先了解现有的产品。此外,我们建议提供自动收集可用报价的发现服务。我们将其与表示发现模型的匹配算法结合起来,并在约束满足问题(CSP)中解决用户给定的约束。最后,我们在运行时操作这个发现模型以对错误作出反应。我们的评估表明,使用基于约束的语言是解决供应商选择问题的可行方法,它有助于克服供应商锁定。
{"title":"A Provider-Agnostic Approach to Multi-cloud Orchestration Using a Constraint Language","authors":"Dani Baur, Daniel Seybold, F. Griesinger, Hynek Masata, Jörg Domaschka","doi":"10.1109/CCGRID.2018.00032","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00032","url":null,"abstract":"Cloud computing and its computing as an utility paradigm provides on-demand resources allowing the seamless adaptation of applications to fluctuating demands. While the Cloud's ongoing commercialisation has lead to a vast provider landscape, vendor lock-in is still a major hindrance. Recent outages demonstrate that relying exclusively on one provider is not sufficient. While existing cloud orchestration tools promise to solve the problems by supporting deployments across multiple cloud providers, they typically rely on provider dependent models forcing prior knowledge of offers and obstructing flexibility in case of errors. We propose a cloud provider-agnostic application and resource description using a constraint language. It allows users to express resource requirements of an application without prior knowledge of existing offers. Additionally, we propose a discovery service automatically collecting available offers. We combine this with a matchmaking algorithm representing the discovery model and the user-given constraints in a constraint satisfaction problem (CSP) that is then solved. Finally, we manipulate this discovery model during runtime to react on errors. Our evaluation shows that using a constraint-based language is a feasible approach to the provider selection problem, and that it helps to overcome vendor lock-in.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132911648","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Enabling Efficient Inter-Node Message Passing and Remote Memory Access Via a uGNI Based Light-Weight Network Substrate for Cray Interconnects 基于uGNI的Cray互连轻量级网络基板实现高效节点间消息传递和远程内存访问
U. Wickramasinghe, A. Lumsdaine
Today's cutting-edge network hardware features extremely low latency and high bandwidth transactions for higher-level communication substrates. The Cray XC/XE family of network fabrics, also known as Cray Aries/Gemini respectively, supports such high-performance remote memory access operations (RMA) and a plethora of transaction modes to optimize communication via lower-level interfaces such as uGNI and DMAPP. However, enabling efficient one-sided communication for higher-level substrates is difficult due to barriers presented by the programming model itself, as well as miscellaneous synchronization bottlenecks at the runtime layers. We present an efficient programming model based on a distributed memory allocator for RMA and a communication substrate based on readers and writers for inter-node message passing and RMA operations. We try to maximize performance by introducing a scalable RMA event notification scheme and synchronization protocols that fully leverage Aries/Gemini fabric. Micro-benchmark results demonstrate that our library outperforms Cray MPI-3.0-based RMA one-sided operations by 1.5X and up to 6X in certain cases and is comparable or improves upon performance on others.
当今尖端的网络硬件具有极低的延迟和高带宽事务,适用于更高级别的通信基板。Cray XC/XE系列网络结构,也分别被称为Cray Aries/Gemini,支持高性能远程内存访问操作(RMA)和大量的事务模式,以优化通过底层接口(如uGNI和DMAPP)的通信。然而,由于编程模型本身存在的障碍以及运行时层的各种同步瓶颈,为更高级别的基板实现有效的单侧通信是困难的。我们提出了一种高效的编程模型,该模型基于RMA的分布式内存分配器和基于读写器的通信基板,用于节点间消息传递和RMA操作。我们试图通过引入可扩展的RMA事件通知方案和充分利用白羊座/双子座结构的同步协议来最大化性能。微基准测试结果表明,我们的库在某些情况下比基于Cray mpi -3.0的RMA单边操作性能高1.5倍,最高可达6倍,并且在其他情况下性能相当或有所提高。
{"title":"Enabling Efficient Inter-Node Message Passing and Remote Memory Access Via a uGNI Based Light-Weight Network Substrate for Cray Interconnects","authors":"U. Wickramasinghe, A. Lumsdaine","doi":"10.1109/CCGRID.2018.00006","DOIUrl":"https://doi.org/10.1109/CCGRID.2018.00006","url":null,"abstract":"Today's cutting-edge network hardware features extremely low latency and high bandwidth transactions for higher-level communication substrates. The Cray XC/XE family of network fabrics, also known as Cray Aries/Gemini respectively, supports such high-performance remote memory access operations (RMA) and a plethora of transaction modes to optimize communication via lower-level interfaces such as uGNI and DMAPP. However, enabling efficient one-sided communication for higher-level substrates is difficult due to barriers presented by the programming model itself, as well as miscellaneous synchronization bottlenecks at the runtime layers. We present an efficient programming model based on a distributed memory allocator for RMA and a communication substrate based on readers and writers for inter-node message passing and RMA operations. We try to maximize performance by introducing a scalable RMA event notification scheme and synchronization protocols that fully leverage Aries/Gemini fabric. Micro-benchmark results demonstrate that our library outperforms Cray MPI-3.0-based RMA one-sided operations by 1.5X and up to 6X in certain cases and is comparable or improves upon performance on others.","PeriodicalId":321027,"journal":{"name":"2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130790216","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2018 18th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1