Aneka云计算环境下的GPU PaaS计算模型

Shashikant Ilager, R. Wankar, Raghavendra Kune, R. Buyya
{"title":"Aneka云计算环境下的GPU PaaS计算模型","authors":"Shashikant Ilager, R. Wankar, Raghavendra Kune, R. Buyya","doi":"10.1201/9780429507670-2","DOIUrl":null,"url":null,"abstract":"Due to the surge in the volume of data generated and rapid advancement in Artificial Intelligence (AI) techniques like machine learning and deep learning, the existing traditional computing models have become inadequate to process an enormous volume of data and the complex application logic for extracting intrinsic information. Computing accelerators such as Graphics processing units (GPUs) have become de facto SIMD computing system for many big data and machine learning applications. On the other hand, the traditional computing model has gradually switched from conventional ownership-based computing to subscription-based cloud computing model. However, the lack of programming models and frameworks to develop cloud-native applications in a seamless manner to utilize both CPU and GPU resources in the cloud has become a bottleneck for rapid application development. To support this application demand for simultaneous heterogeneous resource usage, programming models and new frameworks are needed to manage the underlying resources effectively. Aneka is emerged as a popular PaaS computing model for the development of Cloud applications using multiple programming models like Thread, Task, and MapReduce in a single container .NET platform. Since, Aneka addresses MIMD application development that uses CPU based resources and GPU programming like CUDA is designed for SIMD application development, here, the chapter discusses GPU PaaS computing model for Aneka Clouds for rapid cloud application development for .NET platforms. The popular opensource GPU libraries are utilized and integrated it into the existing Aneka task programming model. The scheduling policies are extended that automatically identify GPU machines and schedule respective tasks accordingly. A case study on image processing is discussed to demonstrate the system, which has been built using PaaS Aneka SDKs and CUDA library.","PeriodicalId":93400,"journal":{"name":"... International Conference on Big Data and Smart Computing. International Conference on Big Data and Smart Computing","volume":"13 1","pages":"19-40"},"PeriodicalIF":0.0000,"publicationDate":"2018-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"GPU PaaS Computation Model in Aneka Cloud Computing Environment\",\"authors\":\"Shashikant Ilager, R. Wankar, Raghavendra Kune, R. Buyya\",\"doi\":\"10.1201/9780429507670-2\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Due to the surge in the volume of data generated and rapid advancement in Artificial Intelligence (AI) techniques like machine learning and deep learning, the existing traditional computing models have become inadequate to process an enormous volume of data and the complex application logic for extracting intrinsic information. Computing accelerators such as Graphics processing units (GPUs) have become de facto SIMD computing system for many big data and machine learning applications. On the other hand, the traditional computing model has gradually switched from conventional ownership-based computing to subscription-based cloud computing model. However, the lack of programming models and frameworks to develop cloud-native applications in a seamless manner to utilize both CPU and GPU resources in the cloud has become a bottleneck for rapid application development. To support this application demand for simultaneous heterogeneous resource usage, programming models and new frameworks are needed to manage the underlying resources effectively. Aneka is emerged as a popular PaaS computing model for the development of Cloud applications using multiple programming models like Thread, Task, and MapReduce in a single container .NET platform. Since, Aneka addresses MIMD application development that uses CPU based resources and GPU programming like CUDA is designed for SIMD application development, here, the chapter discusses GPU PaaS computing model for Aneka Clouds for rapid cloud application development for .NET platforms. The popular opensource GPU libraries are utilized and integrated it into the existing Aneka task programming model. The scheduling policies are extended that automatically identify GPU machines and schedule respective tasks accordingly. A case study on image processing is discussed to demonstrate the system, which has been built using PaaS Aneka SDKs and CUDA library.\",\"PeriodicalId\":93400,\"journal\":{\"name\":\"... International Conference on Big Data and Smart Computing. International Conference on Big Data and Smart Computing\",\"volume\":\"13 1\",\"pages\":\"19-40\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"... International Conference on Big Data and Smart Computing. International Conference on Big Data and Smart Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1201/9780429507670-2\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"... International Conference on Big Data and Smart Computing. International Conference on Big Data and Smart Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1201/9780429507670-2","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

摘要

由于数据量的激增和机器学习、深度学习等人工智能技术的快速发展,现有的传统计算模型已经无法处理海量数据和提取内在信息的复杂应用逻辑。图形处理单元(gpu)等计算加速器已经成为许多大数据和机器学习应用的SIMD计算系统。另一方面,传统的计算模式已经从传统的基于所有权的计算逐渐转向基于订阅的云计算模式。然而,缺乏编程模型和框架来无缝地开发云原生应用程序,以利用云中的CPU和GPU资源,这已经成为快速开发应用程序的瓶颈。为了支持同时使用异构资源的应用程序需求,需要编程模型和新框架来有效地管理底层资源。Aneka是一种流行的PaaS计算模型,用于在单个容器。net平台中使用多线程编程模型(如Thread、Task和MapReduce)开发云应用程序。由于Aneka解决了使用基于CPU资源的MIMD应用程序开发,而GPU编程(如CUDA)是为SIMD应用程序开发而设计的,因此,本章讨论了Aneka云的GPU PaaS计算模型,用于。net平台的快速云应用程序开发。利用流行的开源GPU库并将其集成到现有的Aneka任务编程模型中。扩展了调度策略,自动识别GPU机器并相应地调度相应的任务。最后以图像处理为例,介绍了基于PaaS的Aneka sdk和CUDA库构建的系统。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
GPU PaaS Computation Model in Aneka Cloud Computing Environment
Due to the surge in the volume of data generated and rapid advancement in Artificial Intelligence (AI) techniques like machine learning and deep learning, the existing traditional computing models have become inadequate to process an enormous volume of data and the complex application logic for extracting intrinsic information. Computing accelerators such as Graphics processing units (GPUs) have become de facto SIMD computing system for many big data and machine learning applications. On the other hand, the traditional computing model has gradually switched from conventional ownership-based computing to subscription-based cloud computing model. However, the lack of programming models and frameworks to develop cloud-native applications in a seamless manner to utilize both CPU and GPU resources in the cloud has become a bottleneck for rapid application development. To support this application demand for simultaneous heterogeneous resource usage, programming models and new frameworks are needed to manage the underlying resources effectively. Aneka is emerged as a popular PaaS computing model for the development of Cloud applications using multiple programming models like Thread, Task, and MapReduce in a single container .NET platform. Since, Aneka addresses MIMD application development that uses CPU based resources and GPU programming like CUDA is designed for SIMD application development, here, the chapter discusses GPU PaaS computing model for Aneka Clouds for rapid cloud application development for .NET platforms. The popular opensource GPU libraries are utilized and integrated it into the existing Aneka task programming model. The scheduling policies are extended that automatically identify GPU machines and schedule respective tasks accordingly. A case study on image processing is discussed to demonstrate the system, which has been built using PaaS Aneka SDKs and CUDA library.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Reducing Hadoop 3.x energy consumption through Energy Efficient Ethernet MP-Boost: Minipatch Boosting via Adaptive Feature and Observation Sampling. Minipatch Learning as Implicit Ridge-Like Regularization. Solid-State LiDAR based-SLAM: A Concise Review and Application Anhang B: Begriffe in der qualitativen Inhaltsanalyse
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1