Pub Date : 2025-03-01Epub Date: 2024-12-12DOI: 10.1016/j.peva.2024.102463
Tuhinangshu Choudhury , Gauri Joshi , Weina Wang
Modern machine learning inference systems often host multiple models that can perform the same task with different levels of accuracy and latency. For example, a large model can be more accurate but slow, whereas a smaller and less accurate can be faster in serving inference queries. Amidst the rapid advancements in Large Language Models (LLMs), it is paramount for such systems to strike the best trade-off between latency and accuracy. In this paper, we consider the problem of designing job assignment policies for a multi-server queueing system where servers have heterogeneous rates and accuracies, and our goal is to minimize the expected inference latency while meeting an average accuracy target. Such queueing systems with constraints have been sparsely studied in prior literature to the best of our knowledge. We first identify a lower bound on the minimum achievable latency under any policy that achieves the target accuracy using a linear programming (LP) formulation. Building on the LP solution, we introduce a Randomized-Join-the Idle Queue (R-JIQ) policy, which consistently meets the accuracy target and asymptotically (as system size increases) achieves the optimal latency . However, the R-JIQ policy relies on the knowledge of the arrival rate to solve the LP. To address this limitation, we propose the Prioritize Ordered Pairs (POP) policy that incorporates the concept of ordered pairs of servers into waterfilling to iteratively solve the LP. This allows the POP policy to function without relying on the arrival rate. Experiments suggest that POP performs robustly across different system sizes and load scenarios, achieving near-optimal performance.
{"title":"Job assignment in machine learning inference systems with accuracy constraints","authors":"Tuhinangshu Choudhury , Gauri Joshi , Weina Wang","doi":"10.1016/j.peva.2024.102463","DOIUrl":"10.1016/j.peva.2024.102463","url":null,"abstract":"<div><div>Modern machine learning inference systems often host multiple models that can perform the same task with different levels of accuracy and latency. For example, a large model can be more accurate but slow, whereas a smaller and less accurate can be faster in serving inference queries. Amidst the rapid advancements in Large Language Models (LLMs), it is paramount for such systems to strike the best trade-off between latency and accuracy. In this paper, we consider the problem of designing job assignment policies for a multi-server queueing system where servers have heterogeneous rates and accuracies, and our goal is to minimize the expected inference latency while meeting an average accuracy target. Such queueing systems with constraints have been sparsely studied in prior literature to the best of our knowledge. We first identify a lower bound on the minimum achievable latency under any policy that achieves the target accuracy <span><math><msup><mrow><mi>a</mi></mrow><mrow><mo>∗</mo></mrow></msup></math></span> using a linear programming (LP) formulation. Building on the LP solution, we introduce a Randomized-Join-the Idle Queue (R-JIQ) policy, which consistently meets the accuracy target and asymptotically (as system size increases) achieves the optimal latency <span><math><mrow><msub><mrow><mi>T</mi></mrow><mrow><mtext>LP-LB</mtext></mrow></msub><mrow><mo>(</mo><mi>λ</mi><mo>)</mo></mrow></mrow></math></span>. However, the R-JIQ policy relies on the knowledge of the arrival rate <span><math><mi>λ</mi></math></span> to solve the LP. To address this limitation, we propose the Prioritize Ordered Pairs (POP) policy that incorporates the concept of <em>ordered pairs</em> of servers into waterfilling to iteratively solve the LP. This allows the POP policy to function without relying on the arrival rate. Experiments suggest that POP performs robustly across different system sizes and load scenarios, achieving near-optimal performance.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"167 ","pages":"Article 102463"},"PeriodicalIF":1.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143181843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-01Epub Date: 2024-11-20DOI: 10.1016/j.peva.2024.102459
P. Karan, S. Pradhan
Due to the extensive applications of bulk service vacation queues in manufacturing industries, inventory systems, wireless sensor networks for deducing energy consumption etc., in this article, we analyze the steady-state behavior of an infinite-buffer group arrival bulk service queue with vacation scenario, set-up time and -threshold policy. Here the customers arrive according to the compound Poisson process and the server originates the service process with minimum ‘’ customers and can give service to maximum ‘’ customers at a time. We adopt batch-size-dependent service time as well as queue-length-dependent vacation duration which improve the system’s performance significantly. The -threshold policy is proposed to awaken the server from a vacation/dormant state where the service station starts the set-up procedure after the accumulation of pre-decided ‘’ customers. Using the supplementary variable technique, firstly, we derive the set of system equations in the steady-state. After that, we obtain the bivariate probability generating functions (pgfs) of queue content and size of the departing batch, the queue content and type of vacation taken by the server at vacation completion epoch and also the single pgf of queue content at the end of set-up time. We extract the joint distribution from those generating functions using the roots method and derive a simple algebraic relation between the probabilities at departure and arbitrary epoch. We also provide assorted numerical results to validate our proposed methodology and obtained theoretical results. The impact of the system parameters on the performance measures is presented through tables and graphs. Finally, a cost optimization function is provided for the benefit of system designers.
由于批量服务休假队列在制造业、库存系统、用于推断能源消耗的无线传感器网络等领域的广泛应用,本文分析了具有休假场景、设置时间和 N 个阈值策略的无限缓冲区群到达批量服务队列的稳态行为。在此,客户根据复合泊松过程到达,服务器以最小的 "a "客户启动服务过程,每次可为最大的 "b "客户提供服务。我们采用了与批量大小相关的服务时间和与队列长度相关的休假时间,这大大提高了系统的性能。我们提出了 N 门限策略,用于将服务器从休假/休眠状态唤醒,即服务站在预先确定的 "N "个客户累积后开始设置程序。利用补充变量技术,我们首先推导出稳态下的系统方程组。然后,我们得到了离开批次的队列内容和规模的双变量概率生成函数(pgfs)、服务器在休假结束时的队列内容和休假类型,以及设置时间结束时队列内容的单变量概率生成函数(pgf)。我们使用根法从这些生成函数中提取联合分布,并推导出出发和任意时间点概率之间的简单代数关系。我们还提供了各种数值结果,以验证我们提出的方法和获得的理论结果。我们还通过表格和图表展示了系统参数对性能指标的影响。最后,我们还提供了一个成本优化函数,供系统设计人员参考。
{"title":"Analysis of a queue-length-dependent vacation queue with bulk service, N-policy, set-up time and cost optimization","authors":"P. Karan, S. Pradhan","doi":"10.1016/j.peva.2024.102459","DOIUrl":"10.1016/j.peva.2024.102459","url":null,"abstract":"<div><div>Due to the extensive applications of bulk service vacation queues in manufacturing industries, inventory systems, wireless sensor networks for deducing energy consumption etc., in this article, we analyze the steady-state behavior of an infinite-buffer group arrival bulk service queue with vacation scenario, set-up time and <span><math><mi>N</mi></math></span>-threshold policy. Here the customers arrive according to the compound Poisson process and the server originates the service process with minimum ‘<span><math><mi>a</mi></math></span>’ customers and can give service to maximum ‘<span><math><mi>b</mi></math></span>’ customers at a time. We adopt batch-size-dependent service time as well as queue-length-dependent vacation duration which improve the system’s performance significantly. The <span><math><mi>N</mi></math></span>-threshold policy is proposed to awaken the server from a vacation/dormant state where the service station starts the set-up procedure after the accumulation of pre-decided ‘<span><math><mi>N</mi></math></span>’ customers. Using the supplementary variable technique, firstly, we derive the set of system equations in the steady-state. After that, we obtain the bivariate probability generating functions (pgfs) of queue content and size of the departing batch, the queue content and type of vacation taken by the server at vacation completion epoch and also the single pgf of queue content at the end of set-up time. We extract the joint distribution from those generating functions using the roots method and derive a simple algebraic relation between the probabilities at departure and arbitrary epoch. We also provide assorted numerical results to validate our proposed methodology and obtained theoretical results. The impact of the system parameters on the performance measures is presented through tables and graphs. Finally, a cost optimization function is provided for the benefit of system designers.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"167 ","pages":"Article 102459"},"PeriodicalIF":1.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142723450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-01Epub Date: 2024-12-13DOI: 10.1016/j.peva.2024.102465
Anshul Gandhi , Bo Jiang , Shaolei Ren
{"title":"Editorial: Special issue on Performance Analysis and Evaluation of Systems for Artificial Intelligence","authors":"Anshul Gandhi , Bo Jiang , Shaolei Ren","doi":"10.1016/j.peva.2024.102465","DOIUrl":"10.1016/j.peva.2024.102465","url":null,"abstract":"","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"167 ","pages":"Article 102465"},"PeriodicalIF":1.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143182335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-01Epub Date: 2024-12-18DOI: 10.1016/j.peva.2024.102464
Fabian Michel, Markus Siegle
We study the approximation of a Markov chain on a reduced state space, for both discrete- and continuous-time Markov chains. In this context, we extend the existing theory of formal error bounds for the approximated transient distributions. In the discrete-time setting, we bound the stepwise increment of the error, and in the continuous-time setting, we bound the rate at which the error grows. In addition, the same error bounds can also be applied to bound how far an approximated stationary distribution is from stationarity. As a special case, we consider aggregated (or lumped) Markov chains, where the state space reduction is achieved by partitioning the state space into macro states. Subsequently, we compare the error bounds with relevant concepts from the literature, such as exact and ordinary lumpability, as well as deflatability and aggregatability. These concepts provide stricter than necessary conditions for settings in which the aggregation error is zero. We also present possible algorithms for finding suitable aggregations for which the formal error bounds are low, and we analyze first experiments with these algorithms on a range of different models.
{"title":"Formal error bounds for the state space reduction of Markov chains","authors":"Fabian Michel, Markus Siegle","doi":"10.1016/j.peva.2024.102464","DOIUrl":"10.1016/j.peva.2024.102464","url":null,"abstract":"<div><div>We study the approximation of a Markov chain on a reduced state space, for both discrete- and continuous-time Markov chains. In this context, we extend the existing theory of formal error bounds for the approximated transient distributions. In the discrete-time setting, we bound the stepwise increment of the error, and in the continuous-time setting, we bound the rate at which the error grows. In addition, the same error bounds can also be applied to bound how far an approximated stationary distribution is from stationarity. As a special case, we consider aggregated (or lumped) Markov chains, where the state space reduction is achieved by partitioning the state space into macro states. Subsequently, we compare the error bounds with relevant concepts from the literature, such as exact and ordinary lumpability, as well as deflatability and aggregatability. These concepts provide stricter than necessary conditions for settings in which the aggregation error is zero. We also present possible algorithms for finding suitable aggregations for which the formal error bounds are low, and we analyze first experiments with these algorithms on a range of different models.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"167 ","pages":"Article 102464"},"PeriodicalIF":1.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143181844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-01Epub Date: 2024-11-06DOI: 10.1016/j.peva.2024.102451
Ahsan Ali , Xiaolong Ma , Syed Zawad , Paarijaat Aditya , Istemi Ekin Akkus , Ruichuan Chen , Lei Yang , Feng Yan
In today’s production machine learning (ML) systems, models are continuously trained, improved, and deployed. ML design and training are becoming a continuous workflow of various tasks that have dynamic resource demands. Serverless computing is an emerging cloud paradigm that provides transparent resource management and scaling for users and has the potential to revolutionize the routine of ML design and training. However, hosting modern ML workflows on existing serverless platforms has non-trivial challenges due to their intrinsic design limitations such as stateless nature, limited communication support across function instances, and limited function execution duration. These limitations result in a lack of an overarching view and adaptation mechanism for training dynamics, and an amplification of existing problems in ML workflows.
To address the above challenges, we propose SMLT, an automated, scalable and adaptive serverless framework on public cloud to enable efficient and user-centric ML design and training. SMLT employs an automated and adaptive scheduling mechanism to dynamically optimize the deployment and resource scaling for ML tasks during training. SMLT further enables user-centric ML workflow execution by supporting user-specified training deadline and budget limit. In addition, by providing an end-to-end design, SMLT solves the intrinsic problems in public cloud serverless platforms such as the communication overhead, limited function execution duration, need for repeated initialization, and also provides explicit fault tolerance for ML training. SMLT is open-sourced and compatible with all major ML frameworks. Our experimental evaluation with large, sophisticated modern ML models demonstrates that SMLT outperforms the state-of-the-art VM-based systems and existing public cloud serverless ML training frameworks in both training speed (up to 8) and monetary cost (up to 3).
在当今的生产型机器学习(ML)系统中,模型需要不断训练、改进和部署。ML 的设计和训练正在成为各种任务的连续工作流程,而这些任务都有动态的资源需求。无服务器计算是一种新兴的云计算模式,可为用户提供透明的资源管理和扩展,并有可能彻底改变 ML 设计和训练的常规工作。然而,在现有的无服务器平台上托管现代 ML 工作流面临着非同小可的挑战,原因在于其固有的设计限制,例如无状态特性、跨功能实例的通信支持有限以及功能执行持续时间有限。为了应对上述挑战,我们在公共云上提出了一个自动化、可扩展和自适应的无服务器框架--SMLT,以实现高效和以用户为中心的 ML 设计和训练。SMLT 采用自动化自适应调度机制,在训练过程中动态优化 ML 任务的部署和资源扩展。通过支持用户指定的训练截止日期和预算限制,SMLT 进一步实现了以用户为中心的 ML 工作流执行。此外,通过提供端到端设计,SMLT 解决了公有云无服务器平台的固有问题,如通信开销、有限的函数执行时间、需要重复初始化等,还为 ML 训练提供了显式容错。SMLT 是开源的,兼容所有主要的 ML 框架。我们使用大型、复杂的现代 ML 模型进行的实验评估表明,SMLT 在训练速度(高达 8 倍)和货币成本(高达 3 倍)方面都优于最先进的基于虚拟机的系统和现有的公共云无服务器 ML 训练框架。
{"title":"Enabling scalable and adaptive machine learning training via serverless computing on public cloud","authors":"Ahsan Ali , Xiaolong Ma , Syed Zawad , Paarijaat Aditya , Istemi Ekin Akkus , Ruichuan Chen , Lei Yang , Feng Yan","doi":"10.1016/j.peva.2024.102451","DOIUrl":"10.1016/j.peva.2024.102451","url":null,"abstract":"<div><div>In today’s production machine learning (ML) systems, models are continuously trained, improved, and deployed. ML design and training are becoming a continuous workflow of various tasks that have dynamic resource demands. Serverless computing is an emerging cloud paradigm that provides transparent resource management and scaling for users and has the potential to revolutionize the routine of ML design and training. However, hosting modern ML workflows on existing serverless platforms has non-trivial challenges due to their intrinsic design limitations such as stateless nature, limited communication support across function instances, and limited function execution duration. These limitations result in a lack of an overarching view and adaptation mechanism for training dynamics, and an amplification of existing problems in ML workflows.</div><div>To address the above challenges, we propose <span>SMLT</span>, an automated, scalable and adaptive serverless framework on public cloud to enable efficient and user-centric ML design and training. <span>SMLT</span> employs an automated and adaptive scheduling mechanism to dynamically optimize the deployment and resource scaling for ML tasks during training. <span>SMLT</span> further enables user-centric ML workflow execution by supporting user-specified training deadline and budget limit. In addition, by providing an end-to-end design, <span>SMLT</span> solves the intrinsic problems in public cloud serverless platforms such as the communication overhead, limited function execution duration, need for repeated initialization, and also provides explicit fault tolerance for ML training. <span>SMLT</span> is open-sourced and compatible with all major ML frameworks. Our experimental evaluation with large, sophisticated modern ML models demonstrates that <span>SMLT</span> outperforms the state-of-the-art VM-based systems and existing public cloud serverless ML training frameworks in both training speed (up to 8<span><math><mo>×</mo></math></span>) and monetary cost (up to 3<span><math><mo>×</mo></math></span>).</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"167 ","pages":"Article 102451"},"PeriodicalIF":1.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142704985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-01Epub Date: 2024-11-16DOI: 10.1016/j.peva.2024.102450
Syed Zawad , Xiaolong Ma , Jun Yi , Cheng Li , Minjia Zhang , Lei Yang , Feng Yan , Yuxiong He
Federated Learning (FL) is a new machine learning paradigm that enables training models collaboratively across clients without sharing private data. In FL, data is non-uniformly distributed among clients (i.e., data heterogeneity) and cannot be redistributed nor monitored like in conventional machine learning due to privacy constraints. Such data heterogeneity and privacy requirements bring new challenges for learning hyperparameter optimization as the training dynamics change across clients even within the same training round and they are difficult to be measured due to privacy. The state-of-the-art in hyperparameter customization can greatly improve FL model accuracy but also incur significant computing overheads and power consumption on client devices, and slowdown the training process. To address the prohibitively expensive cost challenge, we explore the possibility of offloading hyperparameter customization to servers. We propose FedCust, a framework that offloads expensive hyperparameter customization cost from the client devices to the central server without violating privacy constraints. Our key discovery is that it is not necessary to do hyperparameter customization for every client, and clients with similar data heterogeneity can use the same hyperparameters to achieve good training performance. We propose heterogeneity measurement metrics for clustering clients into groups such that clients within the same group share hyperparameters. FedCust uses the proxy data from initial model design to emulate different heterogeneity groups and perform hyperparameter customization on the server side without accessing client data nor information. To make the hyperparameter customization scalable, FedCust further employs a Bayesian-strengthened tuner to significantly accelerates the hyperparameter customization speed. Extensive evaluation demonstrates that FedCust achieves up to 7/2/4/4/6% better accuracy than the widely adopted one-size-fits-all approach on popular FL benchmarks FEMNIST, Shakespeare, Cifar100, Cifar10, and Fashion-MNIST respectively, while being scalable and reducing computation, memory, and energy consumption on the client devices, without compromising privacy constraints.
{"title":"FedCust: Offloading hyperparameter customization for federated learning","authors":"Syed Zawad , Xiaolong Ma , Jun Yi , Cheng Li , Minjia Zhang , Lei Yang , Feng Yan , Yuxiong He","doi":"10.1016/j.peva.2024.102450","DOIUrl":"10.1016/j.peva.2024.102450","url":null,"abstract":"<div><div>Federated Learning (FL) is a new machine learning paradigm that enables training models collaboratively across clients without sharing private data. In FL, data is non-uniformly distributed among clients (i.e., data heterogeneity) and cannot be redistributed nor monitored like in conventional machine learning due to privacy constraints. Such data heterogeneity and privacy requirements bring new challenges for learning hyperparameter optimization as the training dynamics change across clients even within the same training round and they are difficult to be measured due to privacy. The state-of-the-art in hyperparameter customization can greatly improve FL model accuracy but also incur significant computing overheads and power consumption on client devices, and slowdown the training process. To address the prohibitively expensive cost challenge, we explore the possibility of offloading hyperparameter customization to servers. We propose <em>FedCust</em>, a framework that offloads expensive hyperparameter customization cost from the client devices to the central server without violating privacy constraints. Our key discovery is that it is not necessary to do hyperparameter customization for every client, and clients with similar data heterogeneity can use the same hyperparameters to achieve good training performance. We propose heterogeneity measurement metrics for clustering clients into groups such that clients within the same group share hyperparameters. <em>FedCust</em> uses the proxy data from initial model design to emulate different heterogeneity groups and perform hyperparameter customization on the server side without accessing client data nor information. To make the hyperparameter customization scalable, <em>FedCust</em> further employs a Bayesian-strengthened tuner to significantly accelerates the hyperparameter customization speed. Extensive evaluation demonstrates that <em>FedCust</em> achieves up to 7/2/4/4/6% better accuracy than the widely adopted one-size-fits-all approach on popular FL benchmarks FEMNIST, Shakespeare, Cifar100, Cifar10, and Fashion-MNIST respectively, while being scalable and reducing computation, memory, and energy consumption on the client devices, without compromising privacy constraints.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"167 ","pages":"Article 102450"},"PeriodicalIF":1.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142704986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-01Epub Date: 2024-12-01DOI: 10.1016/j.peva.2024.102460
Asmad Bin Abdul Razzaque, Andrea Baiocchi
Internet of Things (IoT) is stirring a surge of interest in effective methods for sharing communication channels, with nodes transmitting sporadic, short messages. These messages are often related to control systems that collect sensor data to drive process actuation, such as in industries, autonomous vehicles, and environmental control. Traditional approaches that dominate wireless and cellular communications prove most effective when dealing with a limited number of concurrently active nodes, sending relatively large volumes of data. We address a different scenario where numerous nodes generate and transmit short messages according to non-periodic schedules. In such cases, random multiple access becomes the typical approach for sharing the communication channel. We propose a general modeling framework that enables the investigation of the impact of Successive Interference Cancellation (SIC) on two of the main random access paradigms, namely Slotted ALOHA (SA) and Carrier-Sense Multiple Access (CSMA). The key varying parameter is the target Signal to Interference plus Noise Ratio (SINR) at the receiver, directly tied to the spectral efficiency of the adopted coding and modulation scheme. Two different regimes are highlighted that bring the system to work at relative maxima of the sum-rate. We further investigate the impact of different transmission power settings and imperfect interference cancellation. Leveraging on the insight gained in the saturated node scenario, an adaptive algorithm is defined for the dynamic case, where the number of backlogged nodes varies over time. The numerical results provide evidence of a significant potential for grant-free multiple access, calling for practical algorithms to translate this promise into feasible realizations.
{"title":"Enabling grant-free multiple access through Successive Interference Cancellation","authors":"Asmad Bin Abdul Razzaque, Andrea Baiocchi","doi":"10.1016/j.peva.2024.102460","DOIUrl":"10.1016/j.peva.2024.102460","url":null,"abstract":"<div><div>Internet of Things (IoT) is stirring a surge of interest in effective methods for sharing communication channels, with nodes transmitting sporadic, short messages. These messages are often related to control systems that collect sensor data to drive process actuation, such as in industries, autonomous vehicles, and environmental control. Traditional approaches that dominate wireless and cellular communications prove most effective when dealing with a limited number of concurrently active nodes, sending relatively large volumes of data. We address a different scenario where numerous nodes generate and transmit short messages according to non-periodic schedules. In such cases, random multiple access becomes the typical approach for sharing the communication channel. We propose a general modeling framework that enables the investigation of the impact of Successive Interference Cancellation (SIC) on two of the main random access paradigms, namely Slotted ALOHA (SA) and Carrier-Sense Multiple Access (CSMA). The key varying parameter is the target Signal to Interference plus Noise Ratio (SINR) at the receiver, directly tied to the spectral efficiency of the adopted coding and modulation scheme. Two different regimes are highlighted that bring the system to work at relative maxima of the sum-rate. We further investigate the impact of different transmission power settings and imperfect interference cancellation. Leveraging on the insight gained in the saturated node scenario, an adaptive algorithm is defined for the dynamic case, where the number of backlogged nodes varies over time. The numerical results provide evidence of a significant potential for grant-free multiple access, calling for practical algorithms to translate this promise into feasible realizations.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"167 ","pages":"Article 102460"},"PeriodicalIF":1.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143182242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-01Epub Date: 2024-08-23DOI: 10.1016/j.peva.2024.102440
Mathew L. Wymore, Rohit Sahu, Thomas Ruminski, Vishal Deep, Morgan Ambourn, Gregory Ling, Vishak Narayanan, William Asiedu, Daji Qiao, Henry Duwe
The emerging paradigm of batteryless intermittent sensor networks (BISNs) presents new challenges for researchers of low-power wireless systems and protocols. The nature of these challenges exacerbates the difficulty of evaluating networks of physical sensor nodes, making simulation an even more important component in evaluating performance metrics, such as communication throughput and delay, for BISN designs. To our knowledge, existing simulators and analytical models do not meet the unique needs of BISN research; therefore, we have created a new open-source BISN simulator named Lure. Lure is designed from the ground-up for simulation of batteryless intermittent systems and networks. Written in Python, Lure is powerful, flexible, highly configurable, and supports rapid prototyping of new protocols, systems, and applications, with a low learning curve. In this paper, we present Lure and validate it with experimental data to show that Lure can accurately reflect the reality of BISNs. We then demonstrate the process of applying Lure to research questions in select case studies.
{"title":"Lure: A simulator for networks of batteryless intermittent nodes","authors":"Mathew L. Wymore, Rohit Sahu, Thomas Ruminski, Vishal Deep, Morgan Ambourn, Gregory Ling, Vishak Narayanan, William Asiedu, Daji Qiao, Henry Duwe","doi":"10.1016/j.peva.2024.102440","DOIUrl":"10.1016/j.peva.2024.102440","url":null,"abstract":"<div><p>The emerging paradigm of batteryless intermittent sensor networks (BISNs) presents new challenges for researchers of low-power wireless systems and protocols. The nature of these challenges exacerbates the difficulty of evaluating networks of physical sensor nodes, making simulation an even more important component in evaluating performance metrics, such as communication throughput and delay, for BISN designs. To our knowledge, existing simulators and analytical models do not meet the unique needs of BISN research; therefore, we have created a new open-source BISN simulator named <em>Lure</em>. Lure is designed from the ground-up for simulation of batteryless intermittent systems and networks. Written in Python, Lure is powerful, flexible, highly configurable, and supports rapid prototyping of new protocols, systems, and applications, with a low learning curve. In this paper, we present Lure and validate it with experimental data to show that Lure can accurately reflect the reality of BISNs. We then demonstrate the process of applying Lure to research questions in select case studies.</p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"166 ","pages":"Article 102440"},"PeriodicalIF":1.0,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0166531624000452/pdfft?md5=1c6343234e3ac7dad5efd12075fa6bfd&pid=1-s2.0-S0166531624000452-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142094686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}