首页 > 最新文献

Performance Evaluation最新文献

英文 中文
Simulation modeling of Zoom traffic on a campus network: A case study 校园网上Zoom流量的仿真建模:一个案例研究
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-10-10 DOI: 10.1016/j.peva.2023.102382
Mehdi Karamollahi, Carey Williamson, Martin Arlitt

In this paper, we develop a synthetic workload model for the Zoom network application based on empirical Zoom traffic measurements from a campus network. We then use this model in a simulation study of Zoom network traffic at the campus scale. The simulation results show that hybrid learning places a substantial load on the campus network. Additional simulation experiments investigate the potential benefits of locally-hosted Zoom infrastructure, improved load balancing strategies for Zoom servers, and multicast delivery for Zoom network traffic. The simulation results show that the multicast approach offers the greatest potential benefit for improving Zoom performance on our campus network.

本文基于对某校园网Zoom流量的实证测量,建立了Zoom网络应用的综合工作负载模型。然后,我们使用该模型对校园规模的Zoom网络流量进行了仿真研究。仿真结果表明,混合学习给校园网带来了很大的负荷。另外的仿真实验研究了本地托管的Zoom基础设施、改进的Zoom服务器负载平衡策略以及Zoom网络流量的多播传递的潜在好处。仿真结果表明,组播方法对提高校园网的变焦性能具有最大的潜在效益。
{"title":"Simulation modeling of Zoom traffic on a campus network: A case study","authors":"Mehdi Karamollahi,&nbsp;Carey Williamson,&nbsp;Martin Arlitt","doi":"10.1016/j.peva.2023.102382","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102382","url":null,"abstract":"<div><p>In this paper, we develop a synthetic workload model for the Zoom network application based on empirical Zoom traffic measurements from a campus network. We then use this model in a simulation study of Zoom network traffic at the campus scale. The simulation results show that hybrid learning places a substantial load on the campus network. Additional simulation experiments investigate the potential benefits of locally-hosted Zoom infrastructure, improved load balancing strategies for Zoom servers, and multicast delivery for Zoom network traffic. The simulation results show that the multicast approach offers the greatest potential benefit for improving Zoom performance on our campus network.</p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"162 ","pages":"Article 102382"},"PeriodicalIF":2.2,"publicationDate":"2023-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49874543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Minimizing age of information under arbitrary arrival model with arbitrary packet size 在任意数据包大小的任意到达模型下最小化信息的年龄
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-10-10 DOI: 10.1016/j.peva.2023.102373
Kumar Saurav, Rahul Vaze

We consider a single source–destination pair, where information updates (in short, updates) arrive at the source at arbitrary time instants. For each update, its size, i.e. the service time required for complete transmission to the destination, is also arbitrary. At any time, the source may choose which update to transmit, while incurring transmission cost that is proportional to the duration of transmission. We consider the age of information (AoI) metric that quantifies the staleness of the update (information) at the destination. At any time, AoI is equal to the difference between the current time, and the arrival time of the latest update (at the source) that has been completely transmitted (to the destination). The goal is to find a causal (i.e. online) scheduling policy that minimizes the sum of the AoI and the transmission cost, where the possible decisions at any time are (i) whether to preempt the update under transmission upon arrival of a new update, and (ii) if no update is under transmission, then choose which update to transmit (among the available updates). In this paper, we propose a causal policy called SRPT+ that at each time, (i) preempts the update under transmission if a new update arrives with a smaller size (compared to the remaining size of the update under transmission), and (ii) if no update is under transmission, then from the set of available updates with size less than a threshold (which is a function of the transmission cost and the current AoI), begins to transmit the update for which the ratio of the reduction in AoI upon complete transmission (if not preempted in future) and the remaining size, is maximum. We characterize the performance of SRPT+ using a metric called the competitive ratio, i.e. the ratio of the cost of causal policy and the cost of an optimal offline policy (that knows the entire input in advance), maximized over all possible inputs. We show that the competitive ratio of SRPT+ is at most 5. In the special case when there is no transmission cost, we further show that the competitive ratio of SRPT+ is at most 3.

我们考虑一个单一的源-目标对,其中信息更新(简而言之,更新)在任意时刻到达源。对于每次更新,其大小,即完成传输到目的地所需的服务时间,也是任意的。在任何时候,源可以选择传输哪个更新,同时产生与传输时间成正比的传输成本。我们考虑信息年龄(AoI)度量,它量化了目的地更新(信息)的过时程度。在任何时候,AoI都等于当前时间与已完全传输(到目的地)的最新更新(在源)到达时间之间的差值。目标是找到一个因果(即在线)调度策略,使AoI和传输成本的总和最小化,其中任何时候可能的决策是(i)是否在新更新到达时抢占正在传输的更新,以及(ii)如果没有更新正在传输,则选择传输哪个更新(在可用更新中)。在本文中,我们提出了一种称为SRPT+的因果策略,该策略每次(i)如果新更新以较小的大小到达(与传输下更新的剩余大小相比),则抢占传输下的更新;(ii)如果没有更新正在传输,则从大小小于阈值(这是传输成本和当前AoI的函数)的可用更新集合中,开始传输更新,在完成传输时(如果将来没有被抢占)AoI减少与剩余大小的比率是最大的。我们使用一个称为竞争比的度量来描述SRPT+的性能,即因果策略的成本与最优离线策略(提前知道整个输入)的成本之比,在所有可能的输入中最大化。结果表明,SRPT+的竞争比不超过5。在不存在传输成本的特殊情况下,我们进一步证明SRPT+的竞争比不超过3。
{"title":"Minimizing age of information under arbitrary arrival model with arbitrary packet size","authors":"Kumar Saurav,&nbsp;Rahul Vaze","doi":"10.1016/j.peva.2023.102373","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102373","url":null,"abstract":"<div><p>We consider a single source–destination pair, where information updates (in short, updates) arrive at the source at arbitrary time instants. For each update, its size, i.e. the service time required for complete transmission to the destination, is also arbitrary. At any time, the source may choose which update to transmit, while incurring transmission cost that is proportional to the duration of transmission. We consider the age of information (AoI) metric that quantifies the staleness of the update (information) at the destination. At any time, AoI is equal to the difference between the current time, and the arrival time of the latest update (at the source) that has been completely transmitted (to the destination). The goal is to find a causal (i.e. online) scheduling policy that minimizes the sum of the AoI and the transmission cost, where the possible decisions at any time are (i) whether to preempt the update under transmission upon arrival of a new update, and (ii) if no update is under transmission, then choose which update to transmit (among the available updates). In this paper, we propose a causal policy called SRPT<span><math><msup><mrow></mrow><mrow><mo>+</mo></mrow></msup></math></span> that at each time, (i) preempts the update under transmission if a new update arrives with a smaller size (compared to the remaining size of the update under transmission), and (ii) if no update is under transmission, then from the set of available updates with size less than a threshold (which is a function of the transmission cost and the current AoI), begins to transmit the update for which the ratio of the reduction in AoI upon complete transmission (if not preempted in future) and the remaining size, is maximum. We characterize the performance of SRPT<span><math><msup><mrow></mrow><mrow><mo>+</mo></mrow></msup></math></span><span> using a metric called the competitive ratio, i.e. the ratio of the cost of causal policy and the cost of an optimal offline policy (that knows the entire input in advance), maximized over all possible inputs. We show that the competitive ratio of SRPT</span><span><math><msup><mrow></mrow><mrow><mo>+</mo></mrow></msup></math></span> is at most 5. In the special case when there is no transmission cost, we further show that the competitive ratio of SRPT<span><math><msup><mrow></mrow><mrow><mo>+</mo></mrow></msup></math></span> is at most 3.</p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"162 ","pages":"Article 102373"},"PeriodicalIF":2.2,"publicationDate":"2023-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49874156","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The saturated Multiserver Job Queuing Model with two classes of jobs: Exact and approximate results 具有两类作业的饱和多服务器作业排队模型:精确结果和近似结果
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-10-06 DOI: 10.1016/j.peva.2023.102370
Diletta Olliaro , Marco Ajmone Marsan , Simonetta Balsamo , Andrea Marin

We consider a multiserver queue where jobs request for a varying number of servers for a random service time. The requested number of servers is assigned to each job following a First-In First-Out (FIFO) order. When the number of free servers is not sufficient to accommodate the next job in line, that job and any subsequent jobs in the queue are forced to wait. As a result, not all available servers are allocated to jobs if the next job requires more servers than are currently free. This queuing system is often called a Multiserver Job Queuing Model (MJQM).

In this paper, we study the behavior of a MJQM under saturation, i.e., when the waiting line always contains jobs to be served. We categorize jobs into two classes: the first class consists of jobs that only require one server, while the second class includes jobs that require a larger number of servers. We obtain the system utilization and the throughput of the two job classes for the case in which the number of servers requested by jobs in the second class is equal to the number of available servers, using a simple approach that allows for a general distribution of the service time of jobs in the second class. Hence, we derive the stability condition of the non-saturated MJQM under these assumptions. Additionally, we develop an approximate analysis for the case in which the jobs of the second class require a fraction of the available servers.

Based on analytical and numerical results, we highlight interesting system properties and insights.

我们考虑一个多服务器队列,其中作业在随机服务时间内请求不同数量的服务器。请求的服务器数量按照先进先出(FIFO)顺序分配给每个作业。当可用服务器的数量不足以容纳队列中的下一个作业时,该作业和队列中的任何后续作业都将被迫等待。因此,如果下一个作业需要比当前可用服务器更多的服务器,则不会将所有可用服务器分配给作业。这种排队系统通常被称为多服务器作业队列模型(MJQM)。在本文中,我们研究了MJQM在饱和状态下的行为,即当等待队列总是包含要服务的作业时。我们将作业分为两类:第一类由只需要一台服务器的作业组成,而第二类包括需要更多服务器的作业。对于第二类作业请求的服务器数量等于可用服务器数量的情况,我们使用一种简单的方法来获得两个作业类的系统利用率和吞吐量,该方法允许第二类中作业的服务时间的一般分布。因此,我们导出了在这些假设下非饱和MJQM的稳定性条件。此外,我们对第二类作业需要一小部分可用服务器的情况进行了近似分析。基于分析和数值结果,我们强调了有趣的系统性质和见解。
{"title":"The saturated Multiserver Job Queuing Model with two classes of jobs: Exact and approximate results","authors":"Diletta Olliaro ,&nbsp;Marco Ajmone Marsan ,&nbsp;Simonetta Balsamo ,&nbsp;Andrea Marin","doi":"10.1016/j.peva.2023.102370","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102370","url":null,"abstract":"<div><p><span>We consider a multiserver queue where jobs request for a varying number of servers for a random service time. The requested number of servers is assigned to each job following a First-In First-Out (FIFO) order. When the number of free servers is not sufficient to accommodate the next job in line, that job and any subsequent jobs in the queue are forced to wait. As a result, not all available servers are allocated to jobs if the next job requires more servers than are currently free. This queuing system is often called a </span><span><em>Multiserver Job </em><em>Queuing Model</em></span> (MJQM).</p><p>In this paper, we study the behavior of a MJQM under saturation, i.e., when the waiting line always contains jobs to be served. We categorize jobs into two classes: the first class consists of jobs that only require one server, while the second class includes jobs that require a larger number of servers. We obtain the system utilization and the throughput of the two job classes for the case in which the number of servers requested by jobs in the second class is equal to the number of available servers, using a simple approach that allows for a general distribution of the service time of jobs in the second class. Hence, we derive the stability condition of the non-saturated MJQM under these assumptions. Additionally, we develop an approximate analysis for the case in which the jobs of the second class require a fraction of the available servers.</p><p>Based on analytical and numerical results, we highlight interesting system properties and insights.</p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"162 ","pages":"Article 102370"},"PeriodicalIF":2.2,"publicationDate":"2023-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49874548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Probabilistic indoor tracking of Bluetooth Low-Energy beacons 低功耗蓝牙信标的概率室内跟踪
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-10-05 DOI: 10.1016/j.peva.2023.102374
F. Serhan Daniş , Cem Ersoy , A. Taylan Cemgil

We construct a practical and real-time probabilistic framework for fine target tracking. In our scenario, a Bluetooth Low-Energy (BLE) device navigating in the environment publishes BLE packets that are captured by stationary BLE sensors. The aim is to accurately estimate the live position of the BLE device emitting these packets. The framework is built upon a hidden Markov model (HMM), the components of which are determined with a combination of heuristic and data-driven approaches. In the data-driven part, we rely on the fingerprints formed priorly by extracting received signal strength indicators (RSSI) from the packets. These data are then transformed into probabilistic radio-frequency maps that are used for measuring the likelihood between an RSSI data and a position. The heuristic part involves the movement of the tracked object. Having no access to any inertial information of the object, this movement is modeled with Gaussian densities with variable model parameters that are to be determined heuristically. The practicality of the framework comes from the associated small parameter set used to discretize the components of the HMM. By tuning these parameters, such as the grid size of the area, the mask size and the covariance of the Gaussian; a probabilistic filtering becomes tractable for discrete state spaces. The filtering is then performed by the forward algorithm given the instantaneous sequential RSSI measurements. The performance of the system is evaluated by taking the mean squared errors of the most probable positions at each time step to their corresponding ground-truth positions. We report the statistics of the error distributions and see that we achieve promising results. The approach is also finally evaluated by its runtime and memory usage.

我们构建了一个实用的实时概率框架,用于精细目标跟踪。在我们的场景中,在环境中导航的蓝牙低功耗(BLE)设备发布由固定BLE传感器捕获的BLE数据包。目的是准确估计发射这些数据包的BLE设备的活动位置。该框架建立在隐马尔可夫模型(HMM)之上,隐马尔可夫模型的组成部分采用启发式和数据驱动方法相结合的方法确定。在数据驱动部分,我们依赖于先前通过从数据包中提取接收信号强度指标(RSSI)形成的指纹。然后将这些数据转换为概率射频图,用于测量RSSI数据与位置之间的可能性。启发式部分涉及被跟踪对象的运动。由于无法获得物体的任何惯性信息,这种运动用高斯密度建模,模型参数可变,需要启发式地确定。该框架的实用性来自于用于离散HMM组件的相关小参数集。通过调整这些参数,如网格面积的大小,掩模的大小和高斯的协方差;对于离散状态空间,概率滤波变得易于处理。然后通过给定瞬时连续RSSI测量值的前向算法执行滤波。通过在每个时间步长取最可能位置的均方误差到相应的真值位置来评估系统的性能。我们报告了误差分布的统计数据,并看到我们取得了有希望的结果。最后还根据运行时和内存使用情况对该方法进行评估。
{"title":"Probabilistic indoor tracking of Bluetooth Low-Energy beacons","authors":"F. Serhan Daniş ,&nbsp;Cem Ersoy ,&nbsp;A. Taylan Cemgil","doi":"10.1016/j.peva.2023.102374","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102374","url":null,"abstract":"<div><p><span>We construct a practical and real-time probabilistic framework<span> for fine target tracking. In our scenario, a Bluetooth Low-Energy (BLE) device navigating in the environment publishes BLE packets that are captured by stationary BLE sensors. The aim is to accurately estimate the live position of the BLE device emitting these packets. The framework is built upon a hidden Markov model (HMM), the components of which are determined with a combination of heuristic and data-driven approaches. In the data-driven part, we rely on the fingerprints formed priorly by extracting received signal strength<span> indicators (RSSI) from the packets. These data are then transformed into probabilistic radio-frequency maps that are used for measuring the likelihood between an RSSI data and a position. The heuristic part involves the movement of the tracked object. Having no access to any inertial information of the object, this movement is modeled with Gaussian densities with variable model parameters that are to be determined heuristically. The practicality of the framework comes from the associated small parameter set used to discretize the components of the HMM. By tuning these parameters, such as the grid size of the area, the mask size and the covariance of the Gaussian; a probabilistic filtering becomes tractable for discrete state spaces. The filtering is then performed by the forward algorithm given the instantaneous sequential RSSI measurements. The performance of the system is evaluated by taking the mean squared errors of the most probable positions at each time step to their corresponding ground-truth positions. We report the </span></span></span>statistics of the error distributions and see that we achieve promising results. The approach is also finally evaluated by its runtime and memory usage.</p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"162 ","pages":"Article 102374"},"PeriodicalIF":2.2,"publicationDate":"2023-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49874550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A hardware-independent time estimation method for inference process of convolutional layers on GPU 基于GPU的卷积层推理过程中与硬件无关的时间估计方法
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-09-20 DOI: 10.1016/j.peva.2023.102368
Chengzhen Meng, Hongjun Dai

Nowadays, various AI applications based on Convolutional Neural Networks (CNNs) are widely deployed on GPU-accelerated devices. However, due to the lack of visibility into GPU internal scheduling, accurately modeling the performance of CNN inference tasks or estimating the latency of CNN tasks that are executing or waiting on the GPU is challenging. This hurts the multi-model scheduling on multi-device and CNN real-time inference. Therefore, in this paper, we propose a time estimation method to estimate the forward execution time of a convolutional layer with an arbitrary shape on a GPU. The proposed method divides an explicit General Matrix Multiplication (GEMM) convolution operation into a series of estimatable GPU operations and constructs performance models at the level of sub-operations rather than hardware instructions or entire models. Also, the proposed method can be easily adapted to different hardware devices or underlying algorithm implementations, since it focuses on the variation of execution time relative to the input data scale, without focusing on specific instructions or hardware actions. According to the experiments on four typical CUDA compatible platforms, the proposed method has an average error rate of less than 5% for convolutional layers in some practical CNN models, and has about 8% error rate in estimating GEMM convolution implementations provided by cuDNN library. Experiments show that the proposed method can predict the forward execution time of convolutional layers of arbitrary size in CNN inference tasks on different GPU models.

目前,基于卷积神经网络(cnn)的各种人工智能应用被广泛部署在gpu加速设备上。然而,由于缺乏对GPU内部调度的可见性,准确建模CNN推理任务的性能或估计正在执行或等待GPU的CNN任务的延迟是具有挑战性的。这不利于多设备上的多模型调度和CNN实时推理。因此,在本文中,我们提出了一种时间估计方法来估计任意形状的卷积层在GPU上的前向执行时间。该方法将显式的通用矩阵乘法(General Matrix Multiplication, GEMM)卷积运算分解为一系列可估计的GPU运算,并在子运算层面构建性能模型,而不是硬件指令或整个模型。此外,所提出的方法可以很容易地适应不同的硬件设备或底层算法实现,因为它关注相对于输入数据规模的执行时间的变化,而不关注特定的指令或硬件操作。在四种典型CUDA兼容平台上的实验表明,该方法对一些实际CNN模型的卷积层的平均错误率小于5%,对cuDNN库提供的GEMM卷积实现的估计错误率约为8%。实验表明,该方法可以在不同GPU模型上预测CNN推理任务中任意大小卷积层的前向执行时间。
{"title":"A hardware-independent time estimation method for inference process of convolutional layers on GPU","authors":"Chengzhen Meng,&nbsp;Hongjun Dai","doi":"10.1016/j.peva.2023.102368","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102368","url":null,"abstract":"<div><p><span><span>Nowadays, various AI applications based on </span>Convolutional Neural Networks<span> (CNNs) are widely deployed on GPU-accelerated devices. However, due to the lack of visibility into GPU internal scheduling, accurately modeling the performance of CNN inference tasks or estimating the latency of CNN tasks that are executing or waiting on the GPU is challenging. This hurts the multi-model scheduling on multi-device and CNN real-time inference. Therefore, in this paper, we propose a time estimation<span> method to estimate the forward execution time of a convolutional layer with an arbitrary shape on a GPU. The proposed method divides an explicit </span></span></span><em>General Matrix Multiplication</em><span> (GEMM) convolution<span> operation into a series of estimatable GPU operations and constructs performance models at the level of sub-operations rather than hardware instructions or entire models. Also, the proposed method can be easily adapted to different hardware devices or underlying algorithm implementations, since it focuses on the variation of execution time relative to the input data scale, without focusing on specific instructions or hardware actions. According to the experiments on four typical CUDA compatible platforms, the proposed method has an average error rate of less than 5% for convolutional layers in some practical CNN models, and has about 8% error rate in estimating GEMM convolution implementations provided by cuDNN library. Experiments show that the proposed method can predict the forward execution time of convolutional layers of arbitrary size in CNN inference tasks on different GPU models.</span></span></p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"162 ","pages":"Article 102368"},"PeriodicalIF":2.2,"publicationDate":"2023-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49874549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the regret of online edge service hosting 关于在线边缘服务托管的遗憾
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-09-11 DOI: 10.1016/j.peva.2023.102367
R. Sri Prakash, Nikhil Karamchandani, Sharayu Moharir

We consider the problem of service hosting where a service provider can dynamically rent edge resources via short term contracts to ensure better quality of service to its customers. The service can also be partially hosted at the edge, in which case, customers’ requests can be partially served at the edge. The total cost incurred by the system is modeled as a combination of the rent cost, the service cost incurred due to latency in serving customers, and the fetch cost incurred as a result of the bandwidth used to fetch the code/databases of the service from the cloud servers to host the service at the edge. In this paper, we compare multiple hosting policies with regret as a metric, defined as the difference in the cost incurred by the policy and the optimal policy over some time horizon T. In particular we consider the Retro Renting (RR) and Follow The Perturbed Leader (FTPL) policies proposed in the literature and provide performance guarantees on the regret of these policies. We show that under i.i.d stochastic arrivals, RR policy has linear regret while FTPL policy has constant regret. Next, we propose a variant of FTPL, namely Wait then FTPL (W-FTPL), which also has constant regret while demonstrating much better dependence on the fetch cost. We also show that under adversarial arrivals, RR policy has linear regret while both FTPL and W-FTPL have regret O(T) which is order-optimal.

我们考虑服务托管的问题,服务提供商可以通过短期合同动态租用边缘资源,以确保为客户提供更好的服务质量。服务也可以部分托管在边缘,在这种情况下,客户的请求可以部分在边缘得到服务。系统产生的总成本被建模为租金成本、由于服务客户的延迟而产生的服务成本以及由于用于从云服务器获取服务的代码/数据库以在边缘托管服务的带宽而产生的获取成本的组合。在本文中,我们比较了以后悔为度量标准的多个托管策略,后悔被定义为策略和最优策略在一段时间内产生的成本差异。特别是,我们考虑了文献中提出的回溯租赁(RR)和跟随扰动领导者(FTPL)策略,并为这些策略的后悔提供了性能保证。我们证明了在i.i.d随机到达下,RR策略具有线性后悔,而FTPL策略具有恒定后悔。接下来,我们提出了FTPL的一个变体,即等待然后FTPL(W-FTPL),它也有持续的遗憾,同时表现出对获取成本的更好依赖性。我们还证明了在对抗性到达下,RR策略具有线性后悔,而FTPL和W-FTPL都具有阶最优的后悔O(T)。
{"title":"On the regret of online edge service hosting","authors":"R. Sri Prakash,&nbsp;Nikhil Karamchandani,&nbsp;Sharayu Moharir","doi":"10.1016/j.peva.2023.102367","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102367","url":null,"abstract":"<div><p><span>We consider the problem of service hosting where a service provider can dynamically rent edge resources via short term contracts to ensure better quality of service to its customers. The service can also be partially hosted at the edge, in which case, customers’ requests can be partially served at the edge. The total cost incurred by the system is modeled as a combination of the rent cost, the service cost incurred due to latency in serving customers, and the fetch cost incurred as a result of the bandwidth used to fetch the code/databases of the service from the cloud servers to host the service at the edge. In this paper, we compare multiple hosting policies with regret as a metric, defined as the difference in the cost incurred by the policy and the optimal policy over some time horizon </span><span><math><mi>T</mi></math></span>. In particular we consider the Retro Renting (RR) and Follow The Perturbed Leader (FTPL) policies proposed in the literature and provide performance guarantees on the regret of these policies. We show that under i.i.d stochastic arrivals, RR policy has linear regret while FTPL policy has constant regret. Next, we propose a variant of FTPL, namely Wait then FTPL (W-FTPL), which also has constant regret while demonstrating much better dependence on the fetch cost. We also show that under adversarial arrivals, RR policy has linear regret while both FTPL and W-FTPL have regret <span><math><mrow><mi>O</mi><mrow><mo>(</mo><msqrt><mrow><mi>T</mi></mrow></msqrt><mo>)</mo></mrow></mrow></math></span> which is order-optimal.</p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"162 ","pages":"Article 102367"},"PeriodicalIF":2.2,"publicationDate":"2023-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49874545","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modeling and performance analysis of hybrid systems by queues with setup time 具有建立时间的队列混合系统建模与性能分析
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-09-04 DOI: 10.1016/j.peva.2023.102366
Mitsuki Sato , Kohei Kawamura , Ken’ichi Kawanishi , Tuan Phung-Duc

NFV (Network Functions Virtualization) is a technology to provide network services by applying virtualization. While the virtualization technology provides flexible architecture by itself, a specific environment where physical legacy equipment and virtualized machines coexist is also considered to meet a wide range of requirements from not only users but also service providers. Motivated by such hybrid systems, we propose queueing models with two types of service facilities: legacy servers and virtual machines. Key features to focus on are that while legacy servers are always on standby, virtual machines need setup time to be ready for service because they are shutdown to reduce the power consumption if no jobs are waiting. Keeping in mind delay-sensitive real-time services, we evaluate the performance of the queueing models, in particular, the delay and energy efficiency.

NFV(网络功能虚拟化)是一种通过应用虚拟化来提供网络服务的技术。虽然虚拟化技术本身提供了灵活的体系结构,但物理遗留设备和虚拟化机器共存的特定环境也被认为可以满足用户和服务提供商的广泛需求。受这种混合系统的启发,我们提出了具有两种服务设施的排队模型:遗留服务器和虚拟机。需要关注的关键功能是,虽然传统服务器始终处于待机状态,但虚拟机需要设置时间才能准备好服务,因为如果没有作业等待,它们会关闭以降低功耗。考虑到对延迟敏感的实时服务,我们评估了排队模型的性能,特别是延迟和能量效率。
{"title":"Modeling and performance analysis of hybrid systems by queues with setup time","authors":"Mitsuki Sato ,&nbsp;Kohei Kawamura ,&nbsp;Ken’ichi Kawanishi ,&nbsp;Tuan Phung-Duc","doi":"10.1016/j.peva.2023.102366","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102366","url":null,"abstract":"<div><p>NFV (Network Functions Virtualization) is a technology to provide network services<span> by applying virtualization<span>. While the virtualization technology provides flexible architecture by itself, a specific environment where physical legacy equipment and virtualized machines coexist is also considered to meet a wide range of requirements from not only users but also service providers. Motivated by such hybrid systems, we propose queueing models with two types of service facilities: legacy servers and virtual machines. Key features to focus on are that while legacy servers are always on standby, virtual machines need setup time to be ready for service because they are shutdown to reduce the power consumption if no jobs are waiting. Keeping in mind delay-sensitive real-time services, we evaluate the performance of the queueing models, in particular, the delay and energy efficiency.</span></span></p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"162 ","pages":"Article 102366"},"PeriodicalIF":2.2,"publicationDate":"2023-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49874546","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A loss queueing game for electric vehicle charging performance evaluation 一种用于电动汽车充电性能评价的损失排队博弈
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-09-01 DOI: 10.1016/j.peva.2023.102350
Alix Dupont , Yezekael Hayel , Tania Jiménez , Olivier Beaude , Jean-Baptiste Breal

The demand of electricity at the Charging Stations (CSs) by Electric Vehicle (EV) users is tremendously increasing. However, EV users still face limited resources at the CSs, both in terms of the number of parking spaces equipped with a charging point, and in terms of available power. This paper deals with the choice of a CS among two CSs by the EV users in a competitive environment. The stochastic nature of arrivals and departures at the CSs is modeled by a queueing system. A queueing game is studied where the EV users are the players and choose the CS that gives the highest expected energy received. An approximation of the expected energy received at the CSs is theoretically provided and the quality of this approximation is numerically illustrated and analyzed through simulations. The existence and uniqueness of the equilibrium of the game is proved, and bounds on the Price of Anarchy (PoA) are also provided. Moreover, the model is simulated using a discrete event framework and a sensitivity analysis of the main metrics of the system with respect to the average parking duration and the power sizing coefficient is provided. The results show that the utility of EV users at equilibrium is close to the optimal utility. This study can help a Charging Point Operator (CPO) to design incentives for EV users in order for instance to limit the parking duration so as to improve the social welfare of the EV users.

电动汽车(EV)用户对充电站(CS)的电力需求正在急剧增加。然而,电动汽车用户在CS仍然面临着有限的资源,无论是在配备充电点的停车位数量还是在可用电力方面。本文讨论了在竞争环境中电动汽车用户在两个CS中选择一个CS的问题。CS的到达和离开的随机性质由排队系统建模。研究了一个排队游戏,其中EV用户是玩家,并选择给出最高期望接收能量的CS。理论上提供了CS处接收的预期能量的近似值,并通过仿真对该近似值的质量进行了数值说明和分析。证明了博弈均衡的存在性和唯一性,并给出了无政府价格的界。此外,使用离散事件框架模拟了该模型,并提供了系统的主要指标相对于平均停车持续时间和功率大小系数的灵敏度分析。结果表明,电动汽车用户在均衡状态下的效用接近最优效用。这项研究可以帮助充电点运营商(CPO)为电动汽车用户设计激励措施,例如限制停车时间,以提高电动汽车用户的社会福利。
{"title":"A loss queueing game for electric vehicle charging performance evaluation","authors":"Alix Dupont ,&nbsp;Yezekael Hayel ,&nbsp;Tania Jiménez ,&nbsp;Olivier Beaude ,&nbsp;Jean-Baptiste Breal","doi":"10.1016/j.peva.2023.102350","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102350","url":null,"abstract":"<div><p><span>The demand of electricity at the Charging Stations (CSs) by Electric Vehicle (EV) users is tremendously increasing. However, EV users still face limited resources at the CSs, both in terms of the number of parking spaces equipped with a charging point, and in terms of available power. This paper deals with the choice of a CS among two CSs by the EV users in a competitive environment. The stochastic nature of arrivals and departures at the CSs is modeled by a queueing system. A queueing game is studied where the EV users are the players and choose the CS that gives the highest expected energy received. An </span>approximation of the expected energy received at the CSs is theoretically provided and the quality of this approximation is numerically illustrated and analyzed through simulations. The existence and uniqueness of the equilibrium of the game is proved, and bounds on the Price of Anarchy (PoA) are also provided. Moreover, the model is simulated using a discrete event framework and a sensitivity analysis of the main metrics of the system with respect to the average parking duration and the power sizing coefficient is provided. The results show that the utility of EV users at equilibrium is close to the optimal utility. This study can help a Charging Point Operator (CPO) to design incentives for EV users in order for instance to limit the parking duration so as to improve the social welfare of the EV users.</p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"161 ","pages":"Article 102350"},"PeriodicalIF":2.2,"publicationDate":"2023-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49700929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing recommendations under abandonment risks: Models and algorithms 放弃风险下的推荐优化:模型与算法
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-09-01 DOI: 10.1016/j.peva.2023.102351
Xuchuang Wang , Hong Xie , Pinghui Wang , John C.S. Lui

User abandonment behaviors are quite common in recommendation applications such as online shopping recommendation and news recommendation. To maximize its total “reward” under the risk of user abandonment, the online platform needs to carefully optimize its recommendations for its users. Because inappropriate recommendations can lead to user abandoning the platform, which results in a short learning duration and reduces the cumulative reward. To address this problem, we formulate a new online decision model and propose an algorithmic framework to transfer similar users’ information via parametric estimation, and employ this knowledge to optimize later decisions. The framework’s theoretical guarantees depend on requirements for its transfer learning oracle and online decision oracle. We then design an online learning algorithm consisting of two components that fulfills each corresponding oracle’s requirements. We also conduct extensive experiments to demonstrate our algorithm’s performance.

用户放弃行为在网上购物推荐、新闻推荐等推荐应用中相当常见。为了在用户放弃的风险下最大限度地提高其总“回报”,在线平台需要仔细优化其对用户的推荐。因为不恰当的推荐会导致用户放弃平台,从而导致学习持续时间短,并降低累积奖励。为了解决这个问题,我们建立了一个新的在线决策模型,并提出了一个算法框架,通过参数估计传递相似用户的信息,并利用这些知识来优化后续决策。该框架的理论保证取决于对迁移学习预言机和在线决策预言机的要求。然后,我们设计了一个由两个组件组成的在线学习算法,以满足每个相应oracle的需求。我们还进行了大量的实验来证明我们算法的性能。
{"title":"Optimizing recommendations under abandonment risks: Models and algorithms","authors":"Xuchuang Wang ,&nbsp;Hong Xie ,&nbsp;Pinghui Wang ,&nbsp;John C.S. Lui","doi":"10.1016/j.peva.2023.102351","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102351","url":null,"abstract":"<div><p>User abandonment behaviors are quite common in recommendation applications such as online shopping recommendation and news recommendation. To maximize its total “reward” under the risk of user abandonment, the online platform needs to carefully optimize its recommendations for its users. Because inappropriate recommendations can lead to user abandoning the platform, which results in a short learning duration and reduces the cumulative reward. To address this problem, we formulate a new online decision model and propose an algorithmic framework to <em>transfer similar users’ information</em><span> via parametric estimation, and employ this knowledge to </span><em>optimize later decisions</em><span>. The framework’s theoretical guarantees depend on requirements for its transfer learning oracle and online decision oracle. We then design an online learning algorithm consisting of two components that fulfills each corresponding oracle’s requirements. We also conduct extensive experiments to demonstrate our algorithm’s performance.</span></p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"161 ","pages":"Article 102351"},"PeriodicalIF":2.2,"publicationDate":"2023-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49757519","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CLC: A cross-level program characterization method CLC:一种跨层次的程序表征方法
IF 2.2 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2023-09-01 DOI: 10.1016/j.peva.2023.102354
Li Tang, Scott Pakin

Characterization of program execution plays a key role in performance improvement. There are numerous transformations applied to each step that a program takes on its lowering from source code to a compiler intermediate representation to machine language to microarchitecture-specific execution. The unpredictable benefit of each transformation step could lead a notionally superior algorithm to exhibit inferior performance once actually run, and it can be hard to discern which step in the transformation path contradicted the code developer’s assumptions.

Conventional approaches to program-execution characterization consider the behavior after only a single one of those steps, which limits the information that can be provided to the user. To help address the issue of myopic views of program execution, this paper presents a novel cross-level characterization approach for understanding the behavior of program execution at different levels in the process of writing, compiling, and running a program. We show that this approach provides a richer view of the sources of performance gains and losses and helps identify program execution in a more accurate manner.

程序执行的特征化在性能改进中起着关键作用。程序从源代码到编译器中间表示,从机器语言到微体系结构特定的执行,每一步都有许多转换。每个转换步骤的不可预测的好处可能会导致一个理论上优越的算法在实际运行后表现出较差的性能,而且很难辨别转换路径中的哪一步与代码开发人员的假设相矛盾。程序执行特征化的传统方法只考虑这些步骤中的一个步骤之后的行为,这限制了可以提供给用户的信息。为了帮助解决程序执行的短视问题,本文提出了一种新的跨级别表征方法,用于理解程序在编写、编译和运行过程中不同级别的执行行为。我们表明,这种方法提供了对性能增益和损失来源的更丰富的视图,并有助于以更准确的方式识别程序执行。
{"title":"CLC: A cross-level program characterization method","authors":"Li Tang,&nbsp;Scott Pakin","doi":"10.1016/j.peva.2023.102354","DOIUrl":"https://doi.org/10.1016/j.peva.2023.102354","url":null,"abstract":"<div><p>Characterization of program execution plays a key role in performance improvement. There are numerous transformations applied to each step that a program takes on its lowering from source code to a compiler intermediate representation to machine language to microarchitecture-specific execution. The unpredictable benefit of each transformation step could lead a notionally superior algorithm to exhibit inferior performance once actually run, and it can be hard to discern which step in the transformation path contradicted the code developer’s assumptions.</p><p>Conventional approaches to program-execution characterization consider the behavior after only a single one of those steps, which limits the information that can be provided to the user. To help address the issue of myopic views of program execution, this paper presents a novel cross-level characterization approach for understanding the behavior of program execution at different levels in the process of writing, compiling, and running a program. We show that this approach provides a richer view of the sources of performance gains and losses and helps identify program execution in a more accurate manner.</p></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"161 ","pages":"Article 102354"},"PeriodicalIF":2.2,"publicationDate":"2023-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49701144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Performance Evaluation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1