{"title":"利用能量收集对边缘网络进行分散式 LLM 推断","authors":"Aria Khoshsirat, Giovanni Perin, Michele Rossi","doi":"arxiv-2408.15907","DOIUrl":null,"url":null,"abstract":"Large language models have significantly transformed multiple fields with\ntheir exceptional performance in natural language tasks, but their deployment\nin resource-constrained environments like edge networks presents an ongoing\nchallenge. Decentralized techniques for inference have emerged, distributing\nthe model blocks among multiple devices to improve flexibility and cost\neffectiveness. However, energy limitations remain a significant concern for\nedge devices. We propose a sustainable model for collaborative inference on\ninterconnected, battery-powered edge devices with energy harvesting. A\nsemi-Markov model is developed to describe the states of the devices,\nconsidering processing parameters and average green energy arrivals. This\ninforms the design of scheduling algorithms that aim to minimize device\ndowntimes and maximize network throughput. Through empirical evaluations and\nsimulated runs, we validate the effectiveness of our approach, paving the way\nfor energy-efficient decentralized inference over edge networks.","PeriodicalId":501422,"journal":{"name":"arXiv - CS - Distributed, Parallel, and Cluster Computing","volume":"7 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Decentralized LLM Inference over Edge Networks with Energy Harvesting\",\"authors\":\"Aria Khoshsirat, Giovanni Perin, Michele Rossi\",\"doi\":\"arxiv-2408.15907\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Large language models have significantly transformed multiple fields with\\ntheir exceptional performance in natural language tasks, but their deployment\\nin resource-constrained environments like edge networks presents an ongoing\\nchallenge. Decentralized techniques for inference have emerged, distributing\\nthe model blocks among multiple devices to improve flexibility and cost\\neffectiveness. However, energy limitations remain a significant concern for\\nedge devices. We propose a sustainable model for collaborative inference on\\ninterconnected, battery-powered edge devices with energy harvesting. A\\nsemi-Markov model is developed to describe the states of the devices,\\nconsidering processing parameters and average green energy arrivals. This\\ninforms the design of scheduling algorithms that aim to minimize device\\ndowntimes and maximize network throughput. Through empirical evaluations and\\nsimulated runs, we validate the effectiveness of our approach, paving the way\\nfor energy-efficient decentralized inference over edge networks.\",\"PeriodicalId\":501422,\"journal\":{\"name\":\"arXiv - CS - Distributed, Parallel, and Cluster Computing\",\"volume\":\"7 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Distributed, Parallel, and Cluster Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.15907\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Distributed, Parallel, and Cluster Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.15907","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Decentralized LLM Inference over Edge Networks with Energy Harvesting
Large language models have significantly transformed multiple fields with
their exceptional performance in natural language tasks, but their deployment
in resource-constrained environments like edge networks presents an ongoing
challenge. Decentralized techniques for inference have emerged, distributing
the model blocks among multiple devices to improve flexibility and cost
effectiveness. However, energy limitations remain a significant concern for
edge devices. We propose a sustainable model for collaborative inference on
interconnected, battery-powered edge devices with energy harvesting. A
semi-Markov model is developed to describe the states of the devices,
considering processing parameters and average green energy arrivals. This
informs the design of scheduling algorithms that aim to minimize device
downtimes and maximize network throughput. Through empirical evaluations and
simulated runs, we validate the effectiveness of our approach, paving the way
for energy-efficient decentralized inference over edge networks.