MLPacker: A Unified Software Tool for Packaging and Deploying Atomic and Distributed Analytic Pipelines

Raúl Miñón, Josu Díaz-de-Arcaya, Ana I. Torre-Bastida, Gorka Zárate, Aitor Moreno-Fernandez-de-Leceta
{"title":"MLPacker: A Unified Software Tool for Packaging and Deploying Atomic and Distributed Analytic Pipelines","authors":"Raúl Miñón, Josu Díaz-de-Arcaya, Ana I. Torre-Bastida, Gorka Zárate, Aitor Moreno-Fernandez-de-Leceta","doi":"10.23919/SpliTech55088.2022.9854211","DOIUrl":null,"url":null,"abstract":"In the last years, MLOps (Machine Learning Operations) paradigm is attracting the attention from the community, extrapolating the DevOps (Development and Operations) paradigm to the artificial intelligence (AI) development life-cycle. In this area, some challenges must be addressed to successfully deliver solutions since there are specific nuances when dealing with AI operationalization such as the model packaging or monitoring. Fortunately, interesting and helpful approaches, both from the research community and industry have emerged. However, further research is still necessary to fulfil key gaps. This paper presents a tool, MLPacker, for addressing some of them. Concretely, this tool provides mechanisms to package and deploy analytic pipelines both in REST APIs and in streaming mode. In addition, the analytic pipelines can be deployed atomically (i.e., the whole pipeline in the same machine) or in a distributed fashion (i.e., deploying each stage of the pipeline in distinct machines). In this way, users can take advantage from the cloud continuum paradigm considering edge-fog-cloud computing layers. Finally, the tool is decoupled from the training stage to avoid data scientists the integration of blocks of code in their experiments for the operationalization. Besides the package mode (REST API or streaming), the tool can be configured to perform the deployments in local or in remote machines and by using or not containers. For this aim, this paper describes the gaps this tool addresses, the detailed components and flows supported, as well as an scenario with three different case studies to better explain the research conducted.","PeriodicalId":295373,"journal":{"name":"2022 7th International Conference on Smart and Sustainable Technologies (SpliTech)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 7th International Conference on Smart and Sustainable Technologies (SpliTech)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/SpliTech55088.2022.9854211","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In the last years, MLOps (Machine Learning Operations) paradigm is attracting the attention from the community, extrapolating the DevOps (Development and Operations) paradigm to the artificial intelligence (AI) development life-cycle. In this area, some challenges must be addressed to successfully deliver solutions since there are specific nuances when dealing with AI operationalization such as the model packaging or monitoring. Fortunately, interesting and helpful approaches, both from the research community and industry have emerged. However, further research is still necessary to fulfil key gaps. This paper presents a tool, MLPacker, for addressing some of them. Concretely, this tool provides mechanisms to package and deploy analytic pipelines both in REST APIs and in streaming mode. In addition, the analytic pipelines can be deployed atomically (i.e., the whole pipeline in the same machine) or in a distributed fashion (i.e., deploying each stage of the pipeline in distinct machines). In this way, users can take advantage from the cloud continuum paradigm considering edge-fog-cloud computing layers. Finally, the tool is decoupled from the training stage to avoid data scientists the integration of blocks of code in their experiments for the operationalization. Besides the package mode (REST API or streaming), the tool can be configured to perform the deployments in local or in remote machines and by using or not containers. For this aim, this paper describes the gaps this tool addresses, the detailed components and flows supported, as well as an scenario with three different case studies to better explain the research conducted.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
MLPacker:用于打包和部署原子和分布式分析管道的统一软件工具
在过去的几年里,MLOps(机器学习操作)范式吸引了社区的关注,将DevOps(开发和操作)范式外推到人工智能(AI)开发生命周期。在这个领域,必须解决一些挑战才能成功交付解决方案,因为在处理AI操作化(如模型打包或监控)时存在特定的细微差别。幸运的是,来自研究界和工业界的有趣而有益的方法已经出现。然而,仍需要进一步的研究来填补关键空白。本文提出了一种工具MLPacker来解决其中的一些问题。具体来说,该工具提供了以REST api和流模式打包和部署分析管道的机制。此外,分析管道可以自动部署(即,在同一台机器中部署整个管道)或以分布式方式部署(即,在不同的机器中部署管道的每个阶段)。通过这种方式,用户可以从考虑边缘雾云计算层的云连续体范式中获益。最后,该工具与训练阶段解耦,以避免数据科学家在他们的操作化实验中集成代码块。除了包模式(REST API或流)之外,还可以通过使用或不使用容器将工具配置为在本地或远程机器中执行部署。为了实现这一目标,本文描述了该工具所解决的差距,所支持的详细组件和流程,以及一个包含三个不同案例研究的场景,以更好地解释所进行的研究。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
ZERO ENERGY BUILDINGS: At a Glance Towards real time monitoring of an aeronautical machining process using scalable technologies Predicting TV Viewership with Regression Models Towards Consumer-Oriented Demand Response Systems RFID Thermal Monitoring Sheet (R-TMS) for Skin Temperature Measurements during Superficial Microwave Hyperthermia Treatment
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1