Circumventing Stragglers and Staleness in Distributed CNN using LSTM

A. Ravikumar, Harini Sriraman, Saddikuti Lokesh, Jitendra Sai
{"title":"Circumventing Stragglers and Staleness in Distributed CNN using LSTM","authors":"A. Ravikumar, Harini Sriraman, Saddikuti Lokesh, Jitendra Sai","doi":"10.4108/eetiot.5119","DOIUrl":null,"url":null,"abstract":"INTRODUCTION: Using neural networks for these inherently distributed applications is challenging and time-consuming. There is a crucial need for a framework that supports a distributed deep neural network to yield accurate results at an accelerated time. \nMETHODS: In the proposed framework, any experienced novice user can utilize and execute the neural network models in a distributed manner with the automated hyperparameter tuning feature. In addition, the proposed framework is provided in AWS Sage maker for scaling the distribution and achieving exascale FLOPS. We benchmarked the framework performance by applying it to a medical dataset. \nRESULTS: The maximum performance is achieved with a speedup of 6.59 in 5 nodes. The model encourages expert/ novice neural network users to apply neural network models in the distributed platform and get enhanced results with accelerated training time. There has been a lot of research on how to improve the training time of Convolutional Neural Networks (CNNs) using distributed models, with a particular emphasis on automating the hyperparameter tweaking process. The study shows that training times may be decreased across the board by not just manually tweaking hyperparameters, but also by using L2 regularization, a dropout layer, and ConvLSTM for automatic hyperparameter modification. \nCONCLUSION: The proposed method improved the training speed for model-parallel setups by 1.4% and increased the speed for parallel data by 2.206%. Data-parallel execution achieved a high accuracy of 93.3825%, whereas model-parallel execution achieved a top accuracy of 89.59%.","PeriodicalId":506477,"journal":{"name":"EAI Endorsed Transactions on Internet of Things","volume":"50 21","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"EAI Endorsed Transactions on Internet of Things","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.4108/eetiot.5119","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

INTRODUCTION: Using neural networks for these inherently distributed applications is challenging and time-consuming. There is a crucial need for a framework that supports a distributed deep neural network to yield accurate results at an accelerated time. METHODS: In the proposed framework, any experienced novice user can utilize and execute the neural network models in a distributed manner with the automated hyperparameter tuning feature. In addition, the proposed framework is provided in AWS Sage maker for scaling the distribution and achieving exascale FLOPS. We benchmarked the framework performance by applying it to a medical dataset. RESULTS: The maximum performance is achieved with a speedup of 6.59 in 5 nodes. The model encourages expert/ novice neural network users to apply neural network models in the distributed platform and get enhanced results with accelerated training time. There has been a lot of research on how to improve the training time of Convolutional Neural Networks (CNNs) using distributed models, with a particular emphasis on automating the hyperparameter tweaking process. The study shows that training times may be decreased across the board by not just manually tweaking hyperparameters, but also by using L2 regularization, a dropout layer, and ConvLSTM for automatic hyperparameter modification. CONCLUSION: The proposed method improved the training speed for model-parallel setups by 1.4% and increased the speed for parallel data by 2.206%. Data-parallel execution achieved a high accuracy of 93.3825%, whereas model-parallel execution achieved a top accuracy of 89.59%.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
利用 LSTM 规避分布式 CNN 中的落伍者和停滞现象
简介:将神经网络用于这些固有的分布式应用具有挑战性且耗时。因此亟需一个支持分布式深度神经网络的框架,以便在更短的时间内获得准确的结果。方法:在提议的框架中,任何有经验的新手用户都可以利用自动超参数调整功能,以分布式方式使用和执行神经网络模型。此外,我们还在 AWS Sage maker 中提供了拟议框架,用于扩展分布式系统并实现超大规模 FLOPS。我们将该框架应用于医疗数据集,对其性能进行了基准测试。结果:在 5 个节点上实现了最高性能,速度提高了 6.59 倍。该模型鼓励神经网络专家/新手用户在分布式平台上应用神经网络模型,并通过加快训练时间来获得更好的结果。关于如何利用分布式模型改善卷积神经网络(CNN)的训练时间,已经有很多研究,尤其强调超参数调整过程的自动化。研究表明,不仅可以通过手动调整超参数,还可以通过使用 L2 正则化、剔除层和 ConvLSTM 自动修改超参数来全面缩短训练时间。结论:所提出的方法将模型并行设置的训练速度提高了 1.4%,并行数据的训练速度提高了 2.206%。数据并行执行的准确率高达 93.3825%,而模型并行执行的最高准确率为 89.59%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Robust GAN-Based CNN Model as Generative AI Application for Deepfake Detection Identification of Lithology from Well Log Data Using Machine Learning Crime Prediction using Machine Learning Crime Prediction using Machine Learning Circumventing Stragglers and Staleness in Distributed CNN using LSTM
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1