首页 > 最新文献

Performance Evaluation最新文献

英文 中文
Federated Learning for Collaborative Inference Systems: The case of early exit networks 协同推理系统的联邦学习:早期退出网络的案例
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-12-30 DOI: 10.1016/j.peva.2025.102538
Chuan Xu , Caelin Kaplan , Angelo Rodio , Tareq Si Salem , Giovanni Neglia
In today’s increasingly diverse computing landscape, end devices like sensors and smartphones are progressively equipped with AI models tailored to their local memory and computational constraints. Local inference reduces communication costs and latency; however, these smaller models typically underperform compared to more sophisticated models deployed on edge servers or in the cloud. Collaborative Inference Systems (CISs) address this performance trade-off by enabling smaller devices to offload part of their inference tasks to more capable devices. These systems often deploy hierarchical models that share numerous parameters, exemplified by deep neural networks that utilize strategies like early exits or ordered dropout. In such instances, Federated Learning (FL) may be employed to jointly train the models within a CIS. Yet, traditional training methods have overlooked the operational dynamics of CISs during inference, particularly the potential high heterogeneity in serving rates across the devices within a given CIS. To address this gap, we propose a novel FL approach that explicitly accounts for variations in serving rates within CISs. Our framework not only offers rigorous theoretical guarantees but also surpasses state-of-the-art training algorithms for CISs, especially in scenarios where end devices handle higher inference request rates and where data availability is uneven across devices.
在当今日益多样化的计算环境中,传感器和智能手机等终端设备逐渐配备了根据其本地内存和计算限制量身定制的人工智能模型。本地推理降低了通信成本和延迟;然而,与部署在边缘服务器或云中更复杂的模型相比,这些较小的模型通常表现不佳。协作推理系统(CISs)通过允许较小的设备将其部分推理任务卸载给功能更强大的设备来解决这种性能权衡问题。这些系统通常部署共享众多参数的分层模型,深度神经网络利用早期退出或有序退出等策略就是一个例子。在这种情况下,可以使用联邦学习(FL)来联合训练CIS中的模型。然而,传统的训练方法在推理过程中忽略了CISs的操作动态,特别是在给定的CISs中,设备之间服务率的潜在高度异质性。为了解决这一差距,我们提出了一种新的FL方法,该方法明确地说明了CISs内服务率的变化。我们的框架不仅提供了严格的理论保证,而且还超越了最先进的CISs训练算法,特别是在终端设备处理更高的推理请求率和设备间数据可用性不均匀的情况下。
{"title":"Federated Learning for Collaborative Inference Systems: The case of early exit networks","authors":"Chuan Xu ,&nbsp;Caelin Kaplan ,&nbsp;Angelo Rodio ,&nbsp;Tareq Si Salem ,&nbsp;Giovanni Neglia","doi":"10.1016/j.peva.2025.102538","DOIUrl":"10.1016/j.peva.2025.102538","url":null,"abstract":"<div><div>In today’s increasingly diverse computing landscape, end devices like sensors and smartphones are progressively equipped with AI models tailored to their local memory and computational constraints. Local inference reduces communication costs and latency; however, these smaller models typically underperform compared to more sophisticated models deployed on edge servers or in the cloud. Collaborative Inference Systems (CISs) address this performance trade-off by enabling smaller devices to offload part of their inference tasks to more capable devices. These systems often deploy hierarchical models that share numerous parameters, exemplified by deep neural networks that utilize strategies like early exits or ordered dropout. In such instances, Federated Learning (FL) may be employed to jointly train the models within a CIS. Yet, traditional training methods have overlooked the operational dynamics of CISs during inference, particularly the potential high heterogeneity in serving rates across the devices within a given CIS. To address this gap, we propose a novel FL approach that explicitly accounts for variations in serving rates within CISs. Our framework not only offers rigorous theoretical guarantees but also surpasses state-of-the-art training algorithms for CISs, especially in scenarios where end devices handle higher inference request rates and where data availability is uneven across devices.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"171 ","pages":"Article 102538"},"PeriodicalIF":0.8,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145976292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging task offloading in edge-cloud Computing systems using GI/M(L→K)/1 queueing model with dynamic service rates 利用动态服务速率的GI/M(L→K)/1队列模型在边缘云计算系统中实现任务卸载
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-12-29 DOI: 10.1016/j.peva.2025.102540
Veena Goswami
We consider the offloading of tasks in edge–cloud computing systems using a renewal input modified batch service queue. Tasks are processed using a modified batch service policy with a minimum batch size of L and a maximum batch size of K in an edge–cloud computing system. Bulk services combine several tasks from many Internet of Things devices and offload them to the edge or cloud for concurrent execution. The updated batch service rule allows tasks to be offloaded for variable batch sizes, smaller batches when network circumstances are favorable, and bigger batches when the network is congested to reduce transmission overhead. In addition, if the server has commenced the processing and there are fewer than K tasks, we let the tasks join. Furthermore, the batches’ processing rates are presumed to depend on the batch size. We derive the analytic results for the marginal and joint probability distribution of the number of tasks in the queue/system and with the server. We show the influence of light-tailed and heavy-tailed inter-arrival time distributions on the system model with numerical examples. Dynamic service rates adjust processing speeds at edge or cloud servers based on workload, network latency, and available resources. It reduces latency, balances computational load, and improves system adaptability to changing conditions.
我们考虑了在边缘云计算系统中使用更新输入修改批处理服务队列的任务卸载。在边缘云计算系统中,使用修改后的批处理服务策略处理任务,最小批处理大小为L,最大批处理大小为K。批量服务将来自许多物联网设备的多个任务组合在一起,并将它们卸载到边缘或云以并发执行。更新后的批处理服务规则允许根据不同的批处理大小卸载任务,在网络环境有利时卸载较小的批处理,在网络拥塞时卸载较大的批处理,以减少传输开销。此外,如果服务器已经开始处理并且任务数量少于K,我们将允许这些任务加入。此外,假定批次的处理速率取决于批次大小。我们得到了队列/系统和服务器中任务数的边际概率分布和联合概率分布的分析结果。通过数值算例说明了轻尾和重尾到达时间分布对系统模型的影响。动态服务速率根据工作负载、网络延迟和可用资源调整边缘服务器或云服务器的处理速度。它减少了延迟,平衡了计算负载,提高了系统对变化条件的适应性。
{"title":"Leveraging task offloading in edge-cloud Computing systems using GI/M(L→K)/1 queueing model with dynamic service rates","authors":"Veena Goswami","doi":"10.1016/j.peva.2025.102540","DOIUrl":"10.1016/j.peva.2025.102540","url":null,"abstract":"<div><div>We consider the offloading of tasks in edge–cloud computing systems using a renewal input modified batch service queue. Tasks are processed using a modified batch service policy with a minimum batch size of <span><math><mi>L</mi></math></span> and a maximum batch size of <span><math><mi>K</mi></math></span> in an edge–cloud computing system. Bulk services combine several tasks from many Internet of Things devices and offload them to the edge or cloud for concurrent execution. The updated batch service rule allows tasks to be offloaded for variable batch sizes, smaller batches when network circumstances are favorable, and bigger batches when the network is congested to reduce transmission overhead. In addition, if the server has commenced the processing and there are fewer than <span><math><mi>K</mi></math></span> tasks, we let the tasks join. Furthermore, the batches’ processing rates are presumed to depend on the batch size. We derive the analytic results for the marginal and joint probability distribution of the number of tasks in the queue/system and with the server. We show the influence of light-tailed and heavy-tailed inter-arrival time distributions on the system model with numerical examples. Dynamic service rates adjust processing speeds at edge or cloud servers based on workload, network latency, and available resources. It reduces latency, balances computational load, and improves system adaptability to changing conditions.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"171 ","pages":"Article 102540"},"PeriodicalIF":0.8,"publicationDate":"2025-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145883902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lilou: Resource-aware model-driven latency prediction for GPU-accelerated model serving Lilou: gpu加速模型服务的资源感知模型驱动延迟预测
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-12-27 DOI: 10.1016/j.peva.2025.102539
Qianlin Liang , Haoliang Wang , Prashant Shenoy
As deep learning has been widely used in various application domains, a diversity of GPUs are adopted to accelerate DNN inference workloads and ensure Quality of Service (QoS). Robust prediction of inference latency using GPUs within cloud environments facilitates enhanced efficiency and maintains QoS in resource management solutions, such as consolidation and autoscaling. However, latency prediction is challenging due to the vast heterogeneity in both DNN architectures and GPU capacities.
In this work, we present Lilou, an efficient and accurate latency predicting system for a wide range of DNN inference tasks across diverse GPU resource allocations. Lilou employs two techniques. (i) Lilou represents DNNs as directed acyclic graphs (DAGs), and utilizes a novel graph neural network (GNN) model for edge classification to detect the fusion of operators, also known as kernels. (ii) Lilou identifies the GPU features that significantly impact inference latency and learns a predictor to estimate the latency and type of kernels, which are detected in the preceding step. To evaluate Lilou, we conduct comprehensive experiments across a variety of commercial GPUs commonly utilized in public cloud environments, employing a wide range of popular DNN architectures, including both convolutional neural networks and transformers. Our experiment results show that Lilou is robust to a wide range of DNN architectures and GPU resource allocations. Our novel learning-based method surpasses the state-of-the-art rule-based approach in fusion prediction with an accuracy of 98.26%, laying a solid foundation for end-to-end latency prediction that achieves a MAPE of 8.68%, also outperforming existing benchmarks.
随着深度学习在各个应用领域的广泛应用,采用多种gpu来加速DNN推理工作负载并保证服务质量(QoS)。在云环境中使用gpu对推断延迟进行稳健预测,有助于提高效率,并在资源管理解决方案(如整合和自动缩放)中维护QoS。然而,由于DNN架构和GPU容量的巨大异质性,延迟预测是具有挑战性的。在这项工作中,我们提出了Lilou,这是一个高效准确的延迟预测系统,适用于各种GPU资源分配的DNN推理任务。Lilou采用了两种技术。(i) Lilou将dnn表示为有向无环图(dag),并利用一种新的图神经网络(GNN)模型进行边缘分类,以检测算子(也称为核)的融合。(ii) Lilou识别显著影响推理延迟的GPU特征,并学习一个预测器来估计延迟和内核类型,这些延迟和内核类型在上一步中被检测到。为了评估Lilou,我们在公共云环境中常用的各种商用gpu上进行了全面的实验,采用了各种流行的DNN架构,包括卷积神经网络和变压器。实验结果表明,Lilou对各种DNN架构和GPU资源分配都具有鲁棒性。我们的基于学习的新方法在融合预测方面超过了最先进的基于规则的方法,准确率达到98.26%,为实现端到端延迟预测奠定了坚实的基础,MAPE达到8.68%,也优于现有的基准。
{"title":"Lilou: Resource-aware model-driven latency prediction for GPU-accelerated model serving","authors":"Qianlin Liang ,&nbsp;Haoliang Wang ,&nbsp;Prashant Shenoy","doi":"10.1016/j.peva.2025.102539","DOIUrl":"10.1016/j.peva.2025.102539","url":null,"abstract":"<div><div>As deep learning has been widely used in various application domains, a diversity of GPUs are adopted to accelerate DNN inference workloads and ensure Quality of Service (QoS). Robust prediction of inference latency using GPUs within cloud environments facilitates enhanced efficiency and maintains QoS in resource management solutions, such as consolidation and autoscaling. However, latency prediction is challenging due to the vast heterogeneity in both DNN architectures and GPU capacities.</div><div>In this work, we present <span>Lilou</span>, an efficient and accurate latency predicting system for a wide range of DNN inference tasks across diverse GPU resource allocations. <span>Lilou</span> employs two techniques. (i) <span>Lilou</span> represents DNNs as directed acyclic graphs (DAGs), and utilizes a novel graph neural network (GNN) model for edge classification to detect the fusion of operators, also known as kernels. (ii) <span>Lilou</span> identifies the GPU features that significantly impact inference latency and learns a predictor to estimate the latency and type of kernels, which are detected in the preceding step. To evaluate <span>Lilou</span>, we conduct comprehensive experiments across a variety of commercial GPUs commonly utilized in public cloud environments, employing a wide range of popular DNN architectures, including both convolutional neural networks and transformers. Our experiment results show that <span>Lilou</span> is robust to a wide range of DNN architectures and GPU resource allocations. Our novel learning-based method surpasses the state-of-the-art rule-based approach in fusion prediction with an accuracy of 98.26%, laying a solid foundation for end-to-end latency prediction that achieves a MAPE of 8.68%, also outperforming existing benchmarks.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"171 ","pages":"Article 102539"},"PeriodicalIF":0.8,"publicationDate":"2025-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145924800","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ScaleGS: Scalable distributed framework for large-scale 3D Gaussian splatting with edge communication ScaleGS:用于具有边缘通信的大规模3D高斯飞溅的可扩展分布式框架
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-20 DOI: 10.1016/j.peva.2025.102516
Yong Kou , Jinlong He , Xia Yuan , Dening Luo , Yanci Zhang
3D Gaussian Splatting (3DGS) has recently demonstrated outstanding performance in 3D reconstruction and real-time rendering. However, its scalability to large scenes remains limited by single-GPU memory constraints. We propose ScaleGS, a scalable distributed training framework for large-scale 3DGS with lightweight edge-aware communication. (1) We present a spatial median-guided binary partitioning algorithm that divides the point cloud into balanced, non-overlapping, and spatially contiguous cuboid regions for efficient multi-GPU management. To ensure global view consistency, each GPU independently grows and updates only its local Gaussians, while cross-GPU Gaussians are accessed only for rendering and loss computation. (2) We design a lightweight edge communication strategy to significantly reduce cross-GPU communication overhead. A greedy GPU-Tile remapping algorithm leverages the spatial concentration of Gaussians to confine cross-GPU communication to edge regions, effectively decoupling communication complexity from GPU count, with per-GPU complexity remaining O(1). An optimized all-to-all communication scheme is also introduced to eliminate redundant transmissions. (3) Our framework introduces an adaptive edge-refined load balancing mechanism that periodically monitors GPU workloads and selectively migrates Gaussians between neighboring GPUs to maintain balance and spatial continuity with negligible cost. Evaluations on large-scale 4K scenes show that ScaleGS consistently outperforms state-of-the-art methods, achieving up to 20% faster training and approximately 20% model size reduction on 8 T P40 GPUs without compromising reconstruction quality. Project page: https://aicodeclub.github.io/ScaleGS.
三维高斯溅射(3DGS)技术近年来在三维重建和实时渲染方面表现出优异的性能。然而,它对大型场景的可扩展性仍然受到单gpu内存限制的限制。我们提出了ScaleGS,一个可扩展的分布式训练框架,用于大规模3DGS,具有轻量级边缘感知通信。(1)提出了一种空间中值引导二值分割算法,该算法将点云划分为平衡、不重叠和空间连续的长方体区域,以实现高效的多gpu管理。为了保证全局视图的一致性,每个GPU只独立增长和更新其本地高斯分布,而跨GPU高斯分布仅用于渲染和损耗计算。(2)设计轻量级边缘通信策略,显著降低跨gpu通信开销。贪婪GPU- tile重映射算法利用高斯的空间集中将跨GPU通信限制在边缘区域,有效地将通信复杂度与GPU计数解耦,每个GPU的复杂度保持为0(1)。为了消除冗余传输,采用了一种优化的全对全通信方案。(3)我们的框架引入了一种自适应边缘精细负载平衡机制,该机制定期监控GPU工作负载,并在相邻GPU之间选择性地迁移高斯值,以保持平衡和空间连续性,成本可以忽略不计。对大规模4K场景的评估表明,ScaleGS始终优于最先进的方法,在不影响重建质量的情况下,在8 T P40 gpu上实现了高达20%的训练速度和大约20%的模型尺寸缩小。项目页面:https://aicodeclub.github.io/ScaleGS。
{"title":"ScaleGS: Scalable distributed framework for large-scale 3D Gaussian splatting with edge communication","authors":"Yong Kou ,&nbsp;Jinlong He ,&nbsp;Xia Yuan ,&nbsp;Dening Luo ,&nbsp;Yanci Zhang","doi":"10.1016/j.peva.2025.102516","DOIUrl":"10.1016/j.peva.2025.102516","url":null,"abstract":"<div><div>3D Gaussian Splatting (3DGS) has recently demonstrated outstanding performance in 3D reconstruction and real-time rendering. However, its scalability to large scenes remains limited by single-GPU memory constraints. We propose ScaleGS, a scalable distributed training framework for large-scale 3DGS with lightweight edge-aware communication. (1) We present a spatial median-guided binary partitioning algorithm that divides the point cloud into balanced, non-overlapping, and spatially contiguous cuboid regions for efficient multi-GPU management. To ensure global view consistency, each GPU independently grows and updates only its local Gaussians, while cross-GPU Gaussians are accessed only for rendering and loss computation. (2) We design a lightweight edge communication strategy to significantly reduce cross-GPU communication overhead. A greedy GPU-Tile remapping algorithm leverages the spatial concentration of Gaussians to confine cross-GPU communication to edge regions, effectively decoupling communication complexity from GPU count, with per-GPU complexity remaining <span><math><mrow><mi>O</mi><mrow><mo>(</mo><mn>1</mn><mo>)</mo></mrow></mrow></math></span>. An optimized all-to-all communication scheme is also introduced to eliminate redundant transmissions. (3) Our framework introduces an adaptive edge-refined load balancing mechanism that periodically monitors GPU workloads and selectively migrates Gaussians between neighboring GPUs to maintain balance and spatial continuity with negligible cost. Evaluations on large-scale 4K scenes show that ScaleGS consistently outperforms state-of-the-art methods, achieving up to 20% faster training and approximately 20% model size reduction on 8 T P40 GPUs without compromising reconstruction quality. Project page: <span><span>https://aicodeclub.github.io/ScaleGS</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"171 ","pages":"Article 102516"},"PeriodicalIF":0.8,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145622460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving nonpreemptive multiserver job scheduling with quickswap 使用quickswap改进非抢占式多服务器作业调度
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-17 DOI: 10.1016/j.peva.2025.102525
Zhongrui Chen , Adityo Anggraito , Diletta Olliaro , Andrea Marin , Marco Ajmone Marsan , Benjamin Berg , Isaac Grosof
Modern data center workloads are composed of multiserver jobs, computational jobs that require multiple servers in order to run. A data center can run many multiserver jobs in parallel, as long as it has sufficient resources to meet their individual demands. Multiserver jobs are generally stateful, meaning that job preemptions incur significant overhead from saving and reloading the state associated with running jobs. Hence, most systems try to avoid these costly job preemptions altogether. Given these constraints, a scheduling policy must determine what set of jobs to run in parallel at each moment in time to minimize the mean response time across a stream of arriving jobs. Unfortunately, simple non-preemptive policies such as First-Come First-Served (FCFS) may leave many servers idle, resulting in high mean response times or even system instability. Our goal is to design and analyze non-preemptive scheduling policies for multiserver jobs that maintain high system utilization to achieve low mean response time.
One well-known non-preemptive scheduling policy, Most Servers First (MSF), prioritizes jobs with higher server needs and is known for achieving high resource utilization. However, MSF causes extreme variability in job waiting times, and can perform significantly worse than FCFS in practice. To address this issue, we propose and analyze a class of scheduling policies called Most Servers First with Quickswap (MSFQ) that performs well in a wide variety of cases. MSFQ reduces the variability of job waiting times by periodically granting priority to other jobs in the system. We provide both stability results and an analysis of mean response time under MSFQ to prove that our policy dramatically outperforms MSF in the case where jobs either request one server or all the servers. In more complex cases, we evaluate MSFQ in simulation. We show that, with some additional optimization, variants of the MSFQ policy can greatly outperform MSF and FCFS on real-world multiserver job workloads.
现代数据中心工作负载由多服务器作业组成,这些计算作业需要多台服务器才能运行。数据中心可以并行运行多个多服务器作业,只要它有足够的资源来满足它们各自的需求。多服务器作业通常是有状态的,这意味着作业抢占会因保存和重新加载与运行作业相关的状态而产生大量开销。因此,大多数系统都试图完全避免这些代价高昂的作业抢占。考虑到这些约束,调度策略必须确定在每个时刻并行运行哪一组作业,以最小化到达作业流的平均响应时间。不幸的是,简单的非抢占策略,如先到先得(FCFS)可能会使许多服务器闲置,从而导致高平均响应时间甚至系统不稳定。我们的目标是设计和分析多服务器作业的非抢占调度策略,以保持高系统利用率,从而实现低平均响应时间。一个著名的非抢占式调度策略,大多数服务器优先(MSF),优先考虑服务器需求较高的作业,并以实现高资源利用率而闻名。然而,MSF会导致作业等待时间的极端变化,并且在实践中可能比FCFS表现得更差。为了解决这个问题,我们提出并分析了一类调度策略,称为大多数服务器优先与快速交换(MSFQ),它在各种情况下都表现良好。MSFQ通过定期向系统中的其他作业授予优先级来减少作业等待时间的可变性。我们提供了稳定性结果和MSFQ下的平均响应时间分析,以证明在作业请求一台服务器或所有服务器的情况下,我们的策略显着优于MSF。在更复杂的情况下,我们在模拟中评估MSFQ。我们表明,通过一些额外的优化,MSFQ策略的变体可以在实际的多服务器作业工作负载上大大优于MSF和FCFS。
{"title":"Improving nonpreemptive multiserver job scheduling with quickswap","authors":"Zhongrui Chen ,&nbsp;Adityo Anggraito ,&nbsp;Diletta Olliaro ,&nbsp;Andrea Marin ,&nbsp;Marco Ajmone Marsan ,&nbsp;Benjamin Berg ,&nbsp;Isaac Grosof","doi":"10.1016/j.peva.2025.102525","DOIUrl":"10.1016/j.peva.2025.102525","url":null,"abstract":"<div><div>Modern data center workloads are composed of <em>multiserver jobs</em>, computational jobs that require multiple servers in order to run. A data center can run many multiserver jobs in parallel, as long as it has sufficient resources to meet their individual demands. Multiserver jobs are generally <em>stateful</em>, meaning that job preemptions incur significant overhead from saving and reloading the state associated with running jobs. Hence, most systems try to avoid these costly job preemptions altogether. Given these constraints, a <em>scheduling policy</em> must determine what set of jobs to run in parallel at each moment in time to minimize the mean response time across a stream of arriving jobs. Unfortunately, simple non-preemptive policies such as First-Come First-Served (FCFS) may leave many servers idle, resulting in high mean response times or even system instability. Our goal is to design and analyze non-preemptive scheduling policies for multiserver jobs that maintain high system utilization to achieve low mean response time.</div><div>One well-known non-preemptive scheduling policy, Most Servers First (MSF), prioritizes jobs with higher server needs and is known for achieving high resource utilization. However, MSF causes extreme variability in job waiting times, and can perform significantly worse than FCFS in practice. To address this issue, we propose and analyze a class of scheduling policies called <em>Most Servers First with Quickswap</em> (MSFQ) that performs well in a wide variety of cases. MSFQ reduces the variability of job waiting times by periodically granting priority to other jobs in the system. We provide both stability results and an analysis of mean response time under MSFQ to prove that our policy dramatically outperforms MSF in the case where jobs either request one server or all the servers. In more complex cases, we evaluate MSFQ in simulation. We show that, with some additional optimization, variants of the MSFQ policy can greatly outperform MSF and FCFS on real-world multiserver job workloads.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"171 ","pages":"Article 102525"},"PeriodicalIF":0.8,"publicationDate":"2025-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145691456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A control-theoretic perspective on BBR/CUBIC congestion-control competition BBR/CUBIC拥塞控制竞争的控制理论视角
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-13 DOI: 10.1016/j.peva.2025.102529
Simon Scherrer , Adrian Perrig , Stefan Schmid
To understand the fairness properties of the BBR congestion-control algorithm (CCA), previous research has analyzed BBR behavior with a variety of models. However, previous model-based work suffers from a trade-off between accuracy and interpretability: While dynamic fluid models generate highly accurate predictions through simulation, the causes of their predictions cannot be easily understood. In contrast, steady-state models predict CCA behavior in a manner that is intuitively understandable, but often less accurate. This trade-off is especially consequential when analyzing the competition between BBR and traditional loss-based CCAs, as this competition often suffers from instability, i.e., sending-rate oscillation. Steady-state models cannot predict this instability at all, and fluid-model simulation cannot yield analytical results regarding preconditions and severity of the oscillation.
To overcome this trade-off, we extend the recent dynamic fluid model of BBR by means of control theory. Based on this control-theoretic analysis, we derive quantitative conditions for BBR/CUBIC oscillation, identify network settings that are susceptible to instability, and find that these conditions are frequently satisfied by practical networks. Our analysis illuminates the fairness implications of BBR/CUBIC oscillation, namely by deriving and experimentally validating fairness bounds that reflect the extreme rate distributions during oscillation. In summary, our analysis shows that BBR/CUBIC oscillation is frequent and harms BBR fairness, but can be remedied by means of our control-theoretic framework.
为了理解BBR拥塞控制算法(CCA)的公平性,以往的研究用各种模型分析了BBR行为。然而,以前基于模型的工作在准确性和可解释性之间存在权衡:虽然动态流体模型通过模拟产生高度准确的预测,但其预测的原因不容易理解。相比之下,稳态模型以一种直观可理解的方式预测CCA行为,但通常不太准确。在分析BBR和传统的基于损耗的cca之间的竞争时,这种权衡尤其重要,因为这种竞争经常受到不稳定性的影响,即发送速率振荡。稳态模型根本无法预测这种不稳定性,流体模型模拟也无法得出有关振荡的先决条件和严重程度的分析结果。为了克服这种权衡,我们利用控制理论扩展了最近的BBR动态流体模型。在此控制理论分析的基础上,我们导出了BBR/CUBIC振荡的定量条件,确定了易受不稳定影响的网络设置,并发现这些条件在实际网络中经常得到满足。我们的分析阐明了BBR/CUBIC振荡的公平性含义,即通过推导和实验验证反映振荡期间极端速率分布的公平性界限。综上所述,我们的分析表明,BBR/CUBIC振荡是频繁的,损害了BBR的公平性,但可以通过我们的控制理论框架加以补救。
{"title":"A control-theoretic perspective on BBR/CUBIC congestion-control competition","authors":"Simon Scherrer ,&nbsp;Adrian Perrig ,&nbsp;Stefan Schmid","doi":"10.1016/j.peva.2025.102529","DOIUrl":"10.1016/j.peva.2025.102529","url":null,"abstract":"<div><div>To understand the fairness properties of the BBR congestion-control algorithm (CCA), previous research has analyzed BBR behavior with a variety of models. However, previous model-based work suffers from a trade-off between accuracy and interpretability: While dynamic fluid models generate highly accurate predictions through simulation, the causes of their predictions cannot be easily understood. In contrast, steady-state models predict CCA behavior in a manner that is intuitively understandable, but often less accurate. This trade-off is especially consequential when analyzing the competition between BBR and traditional loss-based CCAs, as this competition often suffers from instability, i.e., sending-rate oscillation. Steady-state models cannot predict this instability at all, and fluid-model simulation cannot yield analytical results regarding preconditions and severity of the oscillation.</div><div>To overcome this trade-off, we extend the recent dynamic fluid model of BBR by means of control theory. Based on this control-theoretic analysis, we derive quantitative conditions for BBR/CUBIC oscillation, identify network settings that are susceptible to instability, and find that these conditions are frequently satisfied by practical networks. Our analysis illuminates the fairness implications of BBR/CUBIC oscillation, namely by deriving and experimentally validating fairness bounds that reflect the extreme rate distributions during oscillation. In summary, our analysis shows that BBR/CUBIC oscillation is frequent and harms BBR fairness, but can be remedied by means of our control-theoretic framework.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"171 ","pages":"Article 102529"},"PeriodicalIF":0.8,"publicationDate":"2025-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145976252","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Content and access networks synergies: Tradeoffs in public and private investments by content providers 内容和接入网络的协同作用:内容提供商在公共和私人投资方面的权衡
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-12 DOI: 10.1016/j.peva.2025.102521
Pranay Agarwal , D. Manjunath
The ubiquity of smartphones has fueled content consumption worldwide, leading to an ever-increasing demand for a better Internet experience. This has necessitated an upgrade of the capacity of the access network. The Internet service providers (ISPs) have been demanding that the content providers (CPs) share the cost of upgrading access network infrastructure. A public investment in the infrastructure of a neutral ISP will boost the profit of the CPs, and hence, seems a rational strategy. A CP can also make a private investment in its infrastructure and boost its profits. In this paper, we study the trade-off between public and private investments by a CP when the decision is made under different types of interaction between them. Specifically, we consider four interaction models between CPs—centralized allocation, cooperative game, non-cooperative game, and a bargaining game—and determine the public and private investment for each model. Via numerical results, we evaluate the impact of different incentive structures on the utility of the CPs. We see that the bargaining game can result in higher public investment than the non-cooperative and centralized models. However, this benefit gets reduced if the CPs are incentivized to invest in private infrastructure.
智能手机的普及推动了全球范围内的内容消费,导致人们对更好的互联网体验的需求不断增长。这就需要对接入网的容量进行升级。互联网服务提供商(isp)一直要求内容提供商(CPs)分担升级接入网基础设施的成本。对中立ISP的基础设施进行公共投资将提高CPs的利润,因此,这似乎是一个理性的策略。CP还可以对其基础设施进行私人投资,从而提高其利润。本文研究了公共投资与私人投资在不同互动类型下决策时的权衡问题。具体来说,我们考虑了集中分配、合作博弈、非合作博弈和议价博弈四种cps交互模型,并确定了每种模型下的公共和私人投资。通过数值结果,我们评估了不同激励结构对CPs效用的影响。我们看到,讨价还价博弈比非合作和集中化模式能带来更高的公共投资。然而,如果鼓励CPs投资私人基础设施,这种好处就会减少。
{"title":"Content and access networks synergies: Tradeoffs in public and private investments by content providers","authors":"Pranay Agarwal ,&nbsp;D. Manjunath","doi":"10.1016/j.peva.2025.102521","DOIUrl":"10.1016/j.peva.2025.102521","url":null,"abstract":"<div><div>The ubiquity of smartphones has fueled content consumption worldwide, leading to an ever-increasing demand for a better Internet experience. This has necessitated an upgrade of the capacity of the access network. The Internet service providers (ISPs) have been demanding that the content providers (CPs) share the cost of upgrading access network infrastructure. A <em>public investment</em> in the infrastructure of a neutral ISP will boost the profit of the CPs, and hence, seems a rational strategy. A CP can also make a <em>private investment</em> in its infrastructure and boost its profits. In this paper, we study the trade-off between public and private investments by a CP when the decision is made under different types of interaction between them. Specifically, we consider four interaction models between CPs—centralized allocation, cooperative game, non-cooperative game, and a bargaining game—and determine the public and private investment for each model. Via numerical results, we evaluate the impact of different incentive structures on the utility of the CPs. We see that the bargaining game can result in higher public investment than the non-cooperative and centralized models. However, this benefit gets reduced if the CPs are incentivized to invest in private infrastructure.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"171 ","pages":"Article 102521"},"PeriodicalIF":0.8,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145579193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Strategic pricing and ranking in recommendation systems with seller competition 考虑卖家竞争的推荐系统策略定价与排名
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-01 DOI: 10.1016/j.peva.2025.102518
Tushar Shankar Walunj , Veeraruna Kavitha , Jayakrishnan Nair , Priyank Agarwal
We study a recommendation system where sellers compete for visibility by strategically offering commissions to a platform that optimally curates a ranked menu of items and their respective prices for each customer. Customers interact sequentially with the menu following a cascade click model, and their purchase decisions are influenced by price sensitivity and positions of various items in the menu. We model the seller-platform interaction as a Stackelberg game with sellers as leaders and consider two different games depending on whether the prices are set by the platform or prefixed by the sellers.
It is complicated to find the optimal policy of the platform in complete generality; hence, we solve the problem in an important asymptotic regime. In fact, both the games coincide in this regime, obtained by decreasing the customer exploration rates γ to zero (in this regime, the customers explore fewer items). Through simulations, we illustrate that the limit game well approximates the original game(s) even for exploration probabilities as high as 0.4 (the differences are around 2.54%). Further, the second game (where the sellers prefix the prices) coincides with the approximate game for all values of γ.
The core contribution of this paper lies in characterizing the equilibrium structure of the limit game. We show that when sellers are of different strengths, the standard Nash equilibrium does not exist due to discontinuities in utilities. We instead establish the existence of a novel equilibrium solution, namely ‘μ-connected equilibrium cycle’ (μ-EC), which captures oscillatory strategic responses at the equilibrium. Unlike the (pure) Nash equilibrium, which defines a fixed point of mutual best responses, this is a set-valued solution concept of connected components. This novel equilibrium concept identifies a Cartesian product set of connected action profiles in the continuous action space that satisfies four important properties: stability against external deviations, no external chains, instability against internal deviations, and minimality. We extend a recently introduced solution concept equilibrium cycle to include stability against measure-zero violations and avoid some topological difficulties to propose μ-EC.
我们研究了一个推荐系统,在这个系统中,卖家通过有策略地向一个平台提供佣金来竞争知名度,该平台为每个客户最佳地策划了一个商品排名菜单及其各自的价格。客户按照级联点击模型顺序与菜单交互,他们的购买决策受到价格敏感性和菜单中各种项目位置的影响。我们将卖家与平台的互动建模为Stackelberg游戏,其中卖家是领导者,并根据价格是由平台设定还是由卖家设定来考虑两种不同的游戏。在完全一般情况下寻找平台的最优策略比较复杂;因此,我们在一个重要的渐近区域内解决了这个问题。事实上,这两款游戏都符合这一机制,即通过将用户探索率γ降低至零而获得(在此机制中,用户探索的道具更少)。通过模拟,我们发现即使勘探概率高达0.4(差异约为2.54%),极限博弈也很接近原始博弈。此外,第二个博弈(卖家在价格前加上前缀)与所有γ值的近似博弈一致。本文的核心贡献在于刻画了极限对策的均衡结构。结果表明,当卖者具有不同的优势时,由于效用的不连续,标准纳什均衡不存在。相反,我们建立了一个新的平衡解的存在性,即“μ-连接的平衡循环”(μ-EC),它捕获了平衡处的振荡策略响应。不像(纯粹的)纳什均衡,它定义了一个相互最佳响应的固定点,这是一个连接组件的集值解决概念。这种新的平衡概念确定了连续作用空间中相互连接的作用轮廓的笛卡尔积集,它满足四个重要性质:抗外部偏差的稳定性、无外部链、抗内部偏差的不稳定性和极小性。我们扩展了最近引入的解决方案概念平衡循环,以包括对测度零违反的稳定性,并避免了一些拓扑困难,提出了μ-EC。
{"title":"Strategic pricing and ranking in recommendation systems with seller competition","authors":"Tushar Shankar Walunj ,&nbsp;Veeraruna Kavitha ,&nbsp;Jayakrishnan Nair ,&nbsp;Priyank Agarwal","doi":"10.1016/j.peva.2025.102518","DOIUrl":"10.1016/j.peva.2025.102518","url":null,"abstract":"<div><div>We study a recommendation system where sellers compete for visibility by strategically offering commissions to a platform that optimally curates a ranked menu of items and their respective prices for each customer. Customers interact sequentially with the menu following a cascade click model, and their purchase decisions are influenced by price sensitivity and positions of various items in the menu. We model the seller-platform interaction as a Stackelberg game with sellers as leaders and consider two different games depending on whether the prices are set by the platform or prefixed by the sellers.</div><div>It is complicated to find the optimal policy of the platform in complete generality; hence, we solve the problem in an important asymptotic regime. In fact, both the games coincide in this regime, obtained by decreasing the customer exploration rates <span><math><mi>γ</mi></math></span> to zero (in this regime, the customers explore fewer items). Through simulations, we illustrate that the limit game well approximates the original game(s) even for exploration probabilities as high as 0.4 (the differences are around 2.54%). Further, the second game (where the sellers prefix the prices) coincides with the approximate game for all values of <span><math><mi>γ</mi></math></span>.</div><div>The core contribution of this paper lies in characterizing the equilibrium structure of the limit game. We show that when sellers are of different strengths, the standard Nash equilibrium does not exist due to discontinuities in utilities. We instead establish the existence of a novel equilibrium solution, namely ‘<span><math><mi>μ</mi></math></span>-connected equilibrium cycle’ (<span><math><mi>μ</mi></math></span>-EC), which captures oscillatory strategic responses at the equilibrium. Unlike the (pure) Nash equilibrium, which defines a fixed point of mutual best responses, this is a set-valued solution concept of connected components. This novel equilibrium concept identifies a Cartesian product set of connected action profiles in the continuous action space that satisfies four important properties: stability against external deviations, no external chains, instability against internal deviations, and minimality. We extend a recently introduced solution concept <em>equilibrium cycle</em> to include stability against measure-zero violations and avoid some topological difficulties to propose <span><math><mi>μ</mi></math></span>-EC.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"170 ","pages":"Article 102518"},"PeriodicalIF":0.8,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145516997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bayesian optimization for dynamic pricing and learning 动态定价与学习的贝叶斯优化
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-01 DOI: 10.1016/j.peva.2025.102519
Anush Anand, Pranav Agrawal, Tejas Bodas
Dynamic pricing is the practice of adjusting the selling price of a product to maximize a firm’s revenue by responding to market demand. The literature typically distinguishes between two settings: infinite inventory, where the firm has unlimited stock and time to sell, and finite inventory, where both inventory and selling horizon are limited. In both cases, the central challenge lies in the fact that the demand function — how sales respond to price — is unknown and must be learned from data. Traditional approaches often assume a specific parametric form for the demand function, enabling the use of reinforcement learning (RL) to identify near-optimal pricing strategies. However, such assumptions may not hold in real-world scenarios, limiting the applicability of these methods.
In this work, we propose a Gaussian Process (GP) based nonparametric approach to dynamic pricing that avoids restrictive modeling assumptions. We treat the demand function as a black-box function of the price and develop pricing algorithms based on Bayesian Optimization (BO)—a sample-efficient method for optimizing unknown functions. We present BO-based algorithms tailored for both infinite and finite inventory settings and provide regret guarantees for both regimes, thereby quantifying the learning efficiency of our methods. Through extensive experiments, we demonstrate that our BO-based methods outperform several state-of-the-art RL algorithms in terms of revenue, while requiring fewer assumptions and offering greater robustness. This highlights Bayesian Optimization as a powerful and practical tool for dynamic pricing in complex, uncertain environments.
动态定价是指根据市场需求调整产品销售价格,使企业收益最大化的做法。文献通常区分两种情况:无限库存,公司有无限的库存和时间来销售;有限库存,库存和销售范围都是有限的。在这两种情况下,核心挑战都在于这样一个事实:需求函数——销售对价格的反应——是未知的,必须从数据中学习。传统方法通常假设需求函数具有特定的参数形式,从而能够使用强化学习(RL)来识别近乎最优的定价策略。然而,这些假设在实际场景中可能不成立,从而限制了这些方法的适用性。在这项工作中,我们提出了一种基于高斯过程(GP)的非参数动态定价方法,避免了限制性建模假设。我们将需求函数视为价格的黑盒函数,并开发了基于贝叶斯优化(BO)的定价算法-一种优化未知函数的样本效率方法。我们提出了针对无限和有限库存设置的基于bo的算法,并为这两种制度提供了后悔保证,从而量化了我们方法的学习效率。通过广泛的实验,我们证明基于bo的方法在收益方面优于几种最先进的强化学习算法,同时需要更少的假设并提供更强的鲁棒性。这突出了贝叶斯优化作为一个强大而实用的工具,在复杂的,不确定的环境中动态定价。
{"title":"Bayesian optimization for dynamic pricing and learning","authors":"Anush Anand,&nbsp;Pranav Agrawal,&nbsp;Tejas Bodas","doi":"10.1016/j.peva.2025.102519","DOIUrl":"10.1016/j.peva.2025.102519","url":null,"abstract":"<div><div>Dynamic pricing is the practice of adjusting the selling price of a product to maximize a firm’s revenue by responding to market demand. The literature typically distinguishes between two settings: infinite inventory, where the firm has unlimited stock and time to sell, and finite inventory, where both inventory and selling horizon are limited. In both cases, the central challenge lies in the fact that the demand function — how sales respond to price — is unknown and must be learned from data. Traditional approaches often assume a specific parametric form for the demand function, enabling the use of reinforcement learning (RL) to identify near-optimal pricing strategies. However, such assumptions may not hold in real-world scenarios, limiting the applicability of these methods.</div><div>In this work, we propose a Gaussian Process (GP) based nonparametric approach to dynamic pricing that avoids restrictive modeling assumptions. We treat the demand function as a black-box function of the price and develop pricing algorithms based on Bayesian Optimization (BO)—a sample-efficient method for optimizing unknown functions. We present BO-based algorithms tailored for both infinite and finite inventory settings and provide regret guarantees for both regimes, thereby quantifying the learning efficiency of our methods. Through extensive experiments, we demonstrate that our BO-based methods outperform several state-of-the-art RL algorithms in terms of revenue, while requiring fewer assumptions and offering greater robustness. This highlights Bayesian Optimization as a powerful and practical tool for dynamic pricing in complex, uncertain environments.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"170 ","pages":"Article 102519"},"PeriodicalIF":0.8,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145568452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparing approximations in the ASIP tandem queue 比较ASIP串联队列中的近似
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-01 DOI: 10.1016/j.peva.2025.102523
Wesley Geelen , Maria Vlasiou , Yaron Yeger
The Asymmetric Inclusion Process (ASIP) models unidirectional transport with particle clustering, yet remains analytically intractable for systems beyond small sizes. To address this, we develop two approximation methods: the replica mean-field (RMF) limit, providing a first-order approximation, and the power series algorithm (PSA), a numerical scheme based on traffic intensity expansions. We evaluate these approximations against Monte Carlo simulations for general systems and prior exact results for homogeneous ASIP systems. Both methods yield accurate estimates, with PSA closely matching simulations for both homogeneous and heterogeneous systems, while RMF performing well for early sites but being slightly impacted downstream or as load increases. These approximations offer practical and computationally efficient alternatives to simulation, enabling detailed performance analysis of ASIP tandem queues where exact solutions are unavailable.
不对称包合过程(ASIP)模拟颗粒聚类的单向输运,但对于小尺寸以上的系统仍然难以分析。为了解决这个问题,我们开发了两种近似方法:复制平均场(RMF)极限,提供一阶近似,以及幂级数算法(PSA),一种基于交通强度展开的数值方案。我们对一般系统的蒙特卡罗模拟和齐次ASIP系统的先前精确结果评估了这些近似。这两种方法都产生了准确的估计,PSA与均匀和非均匀系统的模拟密切匹配,而RMF在早期站点表现良好,但在下游或负载增加时受到轻微影响。这些近似提供了实用且计算效率高的模拟替代方案,可以在无法获得精确解决方案的情况下对ASIP串联队列进行详细的性能分析。
{"title":"Comparing approximations in the ASIP tandem queue","authors":"Wesley Geelen ,&nbsp;Maria Vlasiou ,&nbsp;Yaron Yeger","doi":"10.1016/j.peva.2025.102523","DOIUrl":"10.1016/j.peva.2025.102523","url":null,"abstract":"<div><div>The Asymmetric Inclusion Process (ASIP) models unidirectional transport with particle clustering, yet remains analytically intractable for systems beyond small sizes. To address this, we develop two approximation methods: the replica mean-field (RMF) limit, providing a first-order approximation, and the power series algorithm (PSA), a numerical scheme based on traffic intensity expansions. We evaluate these approximations against Monte Carlo simulations for general systems and prior exact results for homogeneous ASIP systems. Both methods yield accurate estimates, with PSA closely matching simulations for both homogeneous and heterogeneous systems, while RMF performing well for early sites but being slightly impacted downstream or as load increases. These approximations offer practical and computationally efficient alternatives to simulation, enabling detailed performance analysis of ASIP tandem queues where exact solutions are unavailable.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"170 ","pages":"Article 102523"},"PeriodicalIF":0.8,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145466620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Performance Evaluation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1