首页 > 最新文献

Performance Evaluation最新文献

英文 中文
swPredictor: A data-driven performance model for distributed data parallelism training on large-scale HPC clusters swPredictor:一个数据驱动的性能模型,用于大规模高性能计算集群上的分布式数据并行性训练
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-01 DOI: 10.1016/j.peva.2025.102530
Xianyu Zhu , Ruohan Wu , Junshi Chen , Hong An
Given the complexity of heterogeneous architectures and multi-node collaboration, large-scale HPC (High-Performance Computing) clusters pose challenges in resource utilization and performance optimization during distributed data parallelism (DDP) training. Performance modeling aims to identify application bottlenecks and guide algorithm design, but existing performance models rarely consider the impact of system architecture on communication performance or provide a systematic analysis of distributed training. To address these issues, this paper proposes swPredictor, a data-driven performance model devised for accurately predicting the performance of DDP training. First, an original performance dataset is developed based on various communication patterns at runtime to avoid systematic errors. Subsequently, a novel multi-branch module FNO-Inception is proposed, combining FNO (Fourier Neural Operator) layer with Inception structure to simultaneously utilize various frequency features. Finally, by introducing the FNO-Inception module, a novel regression model FI-Net is constructed to fit complex nonlinear relationships. The experimental results demonstrate that FI-Net can accurately predict the performance of DDP training on the Sunway OceanLight supercomputer with an overall MAPE of 0.93%, which outperforms the other baseline models.
考虑到异构架构和多节点协作的复杂性,大规模高性能计算集群在分布式数据并行(DDP)训练过程中对资源利用和性能优化提出了挑战。性能建模的目的是识别应用瓶颈,指导算法设计,但现有的性能模型很少考虑系统架构对通信性能的影响,也很少对分布式训练进行系统分析。为了解决这些问题,本文提出了swPredictor,这是一个数据驱动的性能模型,旨在准确预测DDP训练的性能。首先,在运行时基于各种通信模式开发原始性能数据集,以避免系统错误。随后,提出了一种新的多分支模块FNO-Inception,将FNO(傅里叶神经算子)层与Inception结构相结合,同时利用各种频率特征。最后,通过引入FNO-Inception模块,构造了一个新的拟合复杂非线性关系的回归模型FI-Net。实验结果表明,FI-Net在神威海洋之光超级计算机上能够准确预测DDP训练的性能,总体MAPE为0.93%,优于其他基准模型。
{"title":"swPredictor: A data-driven performance model for distributed data parallelism training on large-scale HPC clusters","authors":"Xianyu Zhu ,&nbsp;Ruohan Wu ,&nbsp;Junshi Chen ,&nbsp;Hong An","doi":"10.1016/j.peva.2025.102530","DOIUrl":"10.1016/j.peva.2025.102530","url":null,"abstract":"<div><div>Given the complexity of heterogeneous architectures and multi-node collaboration, large-scale HPC (High-Performance Computing) clusters pose challenges in resource utilization and performance optimization during distributed data parallelism (DDP) training. Performance modeling aims to identify application bottlenecks and guide algorithm design, but existing performance models rarely consider the impact of system architecture on communication performance or provide a systematic analysis of distributed training. To address these issues, this paper proposes swPredictor, a data-driven performance model devised for accurately predicting the performance of DDP training. First, an original performance dataset is developed based on various communication patterns at runtime to avoid systematic errors. Subsequently, a novel multi-branch module FNO-Inception is proposed, combining FNO (Fourier Neural Operator) layer with Inception structure to simultaneously utilize various frequency features. Finally, by introducing the FNO-Inception module, a novel regression model FI-Net is constructed to fit complex nonlinear relationships. The experimental results demonstrate that FI-Net can accurately predict the performance of DDP training on the Sunway OceanLight supercomputer with an overall MAPE of 0.93%, which outperforms the other baseline models.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"170 ","pages":"Article 102530"},"PeriodicalIF":0.8,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145516999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mitigating massive access with Quasi-Deterministic Transmission: Experiments and stationary analysis 用准确定性传输减少大规模访问:实验和平稳分析
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-01 DOI: 10.1016/j.peva.2025.102512
Jacob Bergquist , Erol Gelenbe , Mohammed Nasereddin , Karl Sigman
The Massive Access Problem arises due to devices that forward packets simultaneously to servers in rapid succession, or by malevolent software in devices that flood network nodes with high-intensity traffic. To protect servers from such events, attack detection (AD) software is installed on servers, and the Quasi-Deterministic Transmission Policy (QDTP) has been proposed to “shape traffic” and protect servers, allowing attack detection to proceed in a timely fashion by delaying some of the incoming packets individually based on their arrival times. QDTP does not cause packet loss, and can be designed so that it does not increase end-to-end packet delay. Starting with measurements taken on an experimental test-bed where the QDPT algorithm is installed on a dedicated processor, which precedes the server itself, we show that QDPT protects the server from attacks by accumulating arriving packets at the input of the QDTP processor, then forwarding them at regular intervals to the server. We compare the behaviour of the server, with and without the use of QDTP, showing the improvement it achieves, provided that its “delay” parameter is correctly selected. We analyze the sample paths associated with QDTP and prove that when its delay parameter is chosen in a specific manner, the end-to-end delay of each packet remains unchanged as compared to an ordinary First-In-First-Out system. An approach based on stationary ergodic processes is developed for the stability conditions. Assuming mutually independent and identically distributed inter-arrival times, service times and QDTP delays, we exhibit the positive recurrent structure of a two-dimensional Markov process and its regeneration points.
大量访问问题是由于设备同时快速连续地向服务器转发数据包,或者设备中的恶意软件以高强度流量淹没网络节点而产生的。为了保护服务器免受此类事件的影响,在服务器上安装了攻击检测(AD)软件,并提出了准确定性传输策略(QDTP)来“塑造流量”和保护服务器,允许攻击检测及时进行,根据它们的到达时间分别延迟一些传入数据包。QDTP不会导致数据包丢失,并且可以设计成不增加端到端数据包延迟。从在实验测试台上进行的测量开始,其中QDPT算法安装在专用处理器上,该处理器先于服务器本身,我们展示了QDPT通过在QDTP处理器的输入处积累到达的数据包,然后定期将它们转发到服务器,从而保护服务器免受攻击。我们比较了服务器的行为,在使用和不使用QDTP的情况下,显示了它所实现的改进,前提是它的“延迟”参数被正确选择。我们分析了与QDTP相关的样本路径,并证明当以特定的方式选择其延迟参数时,与普通的先入先出系统相比,每个数据包的端到端延迟保持不变。针对稳定性条件,提出了一种基于平稳遍历过程的方法。假设到达间隔时间、服务时间和QDTP延迟相互独立且分布相同,我们展示了二维马尔可夫过程及其再生点的正循环结构。
{"title":"Mitigating massive access with Quasi-Deterministic Transmission: Experiments and stationary analysis","authors":"Jacob Bergquist ,&nbsp;Erol Gelenbe ,&nbsp;Mohammed Nasereddin ,&nbsp;Karl Sigman","doi":"10.1016/j.peva.2025.102512","DOIUrl":"10.1016/j.peva.2025.102512","url":null,"abstract":"<div><div>The Massive Access Problem arises due to devices that forward packets simultaneously to servers in rapid succession, or by malevolent software in devices that flood network nodes with high-intensity traffic. To protect servers from such events, attack detection (AD) software is installed on servers, and the Quasi-Deterministic Transmission Policy (QDTP) has been proposed to “shape traffic” and protect servers, allowing attack detection to proceed in a timely fashion by delaying some of the incoming packets individually based on their arrival times. QDTP does not cause packet loss, and can be designed so that it does not increase end-to-end packet delay. Starting with measurements taken on an experimental test-bed where the QDPT algorithm is installed on a dedicated processor, which precedes the server itself, we show that QDPT protects the server from attacks by accumulating arriving packets at the input of the QDTP processor, then forwarding them at regular intervals to the server. We compare the behaviour of the server, with and without the use of QDTP, showing the improvement it achieves, provided that its “delay” parameter is correctly selected. We analyze the sample paths associated with QDTP and prove that when its delay parameter is chosen in a specific manner, the end-to-end delay of each packet remains unchanged as compared to an ordinary First-In-First-Out system. An approach based on stationary ergodic processes is developed for the stability conditions. Assuming mutually independent and identically distributed inter-arrival times, service times and QDTP delays, we exhibit the positive recurrent structure of a two-dimensional Markov process and its regeneration points.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"170 ","pages":"Article 102512"},"PeriodicalIF":0.8,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145466621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CommonSense: Efficient Set Intersection (SetX) protocol based on compressed sensing CommonSense:基于压缩感知的高效集交集(SetX)协议
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-11-01 DOI: 10.1016/j.peva.2025.102520
Jingfan Meng, Tianji Yang, Jun Xu
Set reconciliation (SetR) is an important research problem that has been studied for over two decades. In this problem, two large sets A and B of objects (tokens, files, records, etc.) are stored respectively at two different network-connected hosts, which we name Alice and Bob respectively. Alice and Bob need to communicate with each other to learn the set union AB (which then becomes their reconciled state), at low communication and computation costs. In this work, we study a different problem intricately related to SetR: Alice and Bob collaboratively compute AB. We call this problem SetX (set intersection). Although SetX is just as important as SetR, it has never been properly studied in its own right. Rather, there is an unspoken perception by the research community that SetR and SetX are equally difficult (in costs), and hence “roughly equivalent.” Our first contribution is to show that SetX is fundamentally a much “cheaper” problem than SetR, debunking this long-standing perception. Our second contribution is to develop a novel SetX solution, the communication cost of which handily beats the information-theoretic lower bound of SetR. This protocol is based on the idea of compressed sensing (CS), which we describe here only for the special case of AB (We do have a more sophisticated protocol for the general case). Our protocol is for Alice to encode A into a CS sketch M1A and send it to Bob, where M is a CS matrix with l rows and 1A is the binary vector representation of A. Our key innovation here is to make l (the sketch size) just large enough (for the sketch) to summarize BA (what Alice misses). In contrast, in existing protocols l needs to be large enough to summarize A (what Alice knows), which is typically much larger in cardinality. Our third contribution is to design a CS matrix M that is both “friendly” to (the performance of) applications and “compliant” with CS theory.
集调和(SetR)是一个重要的研究问题,已经被研究了二十多年。在这个问题中,对象(令牌,文件,记录等)的两个大集合A和B分别存储在两个不同的网络连接的主机上,我们分别将其命名为Alice和Bob。Alice和Bob需要相互通信以学习集合并集A∪B(然后成为他们的协调状态),通信和计算成本很低。在这项工作中,我们研究了一个与SetR复杂相关的不同问题:Alice和Bob协同计算a∩B。我们称这个问题为SetX(集合交集)。尽管SetX和SetR一样重要,但它本身从未得到过适当的研究。相反,研究界有一种不言而喻的看法,即SetR和SetX同样困难(在成本上),因此“大致相当”。我们的第一个贡献是表明SetX从根本上来说是一个比SetR“便宜”得多的问题,揭穿了这种长期存在的看法。我们的第二个贡献是开发了一种新的SetX解决方案,其通信成本轻松地超过了SetR的信息论下界。该协议基于压缩感知(CS)的思想,我们在此仅对A≥B的特殊情况进行描述(对于一般情况,我们确实有更复杂的协议)。我们的协议是让Alice将A编码为CS草图M1A并将其发送给Bob,其中M是一个有l行的CS矩阵,1A是A的二进制向量表示。我们这里的关键创新是使l(草图大小)足够大(对于草图)来总结B∈A (Alice遗漏的内容)。相比之下,在现有协议中,l需要足够大来总结A (Alice所知道的),而A的基数通常要大得多。我们的第三个贡献是设计一个CS矩阵M,它对应用程序(性能)既“友好”,又“符合”CS理论。
{"title":"CommonSense: Efficient Set Intersection (SetX) protocol based on compressed sensing","authors":"Jingfan Meng,&nbsp;Tianji Yang,&nbsp;Jun Xu","doi":"10.1016/j.peva.2025.102520","DOIUrl":"10.1016/j.peva.2025.102520","url":null,"abstract":"<div><div>Set reconciliation (SetR) is an important research problem that has been studied for over two decades. In this problem, two large sets <span><math><mi>A</mi></math></span> and <span><math><mi>B</mi></math></span> of objects (tokens, files, records, etc.) are stored respectively at two different network-connected hosts, which we name Alice and Bob respectively. Alice and Bob need to communicate with each other to learn the set union <span><math><mrow><mi>A</mi><mo>∪</mo><mi>B</mi></mrow></math></span> (which then becomes their reconciled state), at low communication and computation costs. In this work, we study a different problem intricately related to SetR: Alice and Bob collaboratively compute <span><math><mrow><mi>A</mi><mo>∩</mo><mi>B</mi></mrow></math></span>. We call this problem SetX (set intersection). Although SetX is just as important as SetR, it has never been properly studied in its own right. Rather, there is an unspoken perception by the research community that SetR and SetX are equally difficult (in costs), and hence “roughly equivalent.” Our first contribution is to show that SetX is fundamentally a much “cheaper” problem than SetR, debunking this long-standing perception. Our second contribution is to develop a novel SetX solution, the communication cost of which handily beats the information-theoretic lower bound of SetR. This protocol is based on the idea of compressed sensing (CS), which we describe here only for the special case of <span><math><mrow><mi>A</mi><mo>⊆</mo><mi>B</mi></mrow></math></span> (We do have a more sophisticated protocol for the general case). Our protocol is for Alice to encode <span><math><mi>A</mi></math></span> into a CS sketch <span><math><mrow><mi>M</mi><msub><mrow><mi>1</mi></mrow><mrow><mi>A</mi></mrow></msub></mrow></math></span> and send it to Bob, where <span><math><mi>M</mi></math></span> is a CS matrix with <span><math><mi>l</mi></math></span> rows and <span><math><msub><mrow><mi>1</mi></mrow><mrow><mi>A</mi></mrow></msub></math></span> is the binary vector representation of <span><math><mi>A</mi></math></span>. Our key innovation here is to make <span><math><mi>l</mi></math></span> (the sketch size) just large enough (for the sketch) to summarize <span><math><mrow><mi>B</mi><mo>∖</mo><mi>A</mi></mrow></math></span> (what Alice misses). In contrast, in existing protocols <span><math><mi>l</mi></math></span> needs to be large enough to summarize <span><math><mi>A</mi></math></span> (what Alice knows), which is typically much larger in cardinality. Our third contribution is to design a CS matrix <span><math><mi>M</mi></math></span> that is both “friendly” to (the performance of) applications and “compliant” with CS theory.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"170 ","pages":"Article 102520"},"PeriodicalIF":0.8,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145517051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
User equilibria in heterogeneous discriminatory processor sharing queues 异构歧视性处理器共享队列中的用户均衡
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-09-24 DOI: 10.1016/j.peva.2025.102510
Dieter Fiems , Balakrishna J. Prabhu
We consider a strategic routing game for a two-class discriminatory processor-sharing queue with an additional cost for joining the premium class. We show that, depending on the specific parameters of the system, various equilibria can coexist, including equilibria where the queueing system is not ergodic for the equilibrium traffic split. We also investigate how the server can select the priority of the classes and the fees charged to the customers to maximise its revenue. We then investigate learning strategies that converge to particular equilibria. Finally, we study how the elasticity of the traffic demand affects the equilibrium solutions.
我们考虑了一个两类歧视性处理器共享队列的策略路由博弈,该队列具有加入优质类的额外成本。我们证明,根据系统的特定参数,各种均衡可以共存,包括排队系统对于均衡流量分割不遍历的均衡。我们还研究了服务器如何选择课程的优先级和向客户收取的费用,以使其收入最大化。然后我们研究收敛于特定均衡的学习策略。最后,研究了交通需求弹性对均衡解的影响。
{"title":"User equilibria in heterogeneous discriminatory processor sharing queues","authors":"Dieter Fiems ,&nbsp;Balakrishna J. Prabhu","doi":"10.1016/j.peva.2025.102510","DOIUrl":"10.1016/j.peva.2025.102510","url":null,"abstract":"<div><div>We consider a strategic routing game for a two-class discriminatory processor-sharing queue with an additional cost for joining the premium class. We show that, depending on the specific parameters of the system, various equilibria can coexist, including equilibria where the queueing system is not ergodic for the equilibrium traffic split. We also investigate how the server can select the priority of the classes and the fees charged to the customers to maximise its revenue. We then investigate learning strategies that converge to particular equilibria. Finally, we study how the elasticity of the traffic demand affects the equilibrium solutions.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"170 ","pages":"Article 102510"},"PeriodicalIF":0.8,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145159602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Response time in a pair of processor sharing queues with Join-the-Shortest-Queue scheduling 具有最短队列加入调度的一对处理器共享队列中的响应时间
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-09-02 DOI: 10.1016/j.peva.2025.102509
Julianna Bor, Peter G. Harrison
Join-the-Shortest-Queue (JSQ) is the scheduling policy of choice for many network providers, cloud servers, and traffic management systems, where individual queues are served under the processor sharing (PS) queueing discipline. A numerical solution for the response time distribution in two parallel PS queues with JSQ scheduling is derived for the first time. Using the generating function method, two partial differential equations (PDEs) are obtained corresponding to conditional response times, where the conditioning is on a particular traced task joining the first or the second queue. These PDEs are functional equations that contain partial generating functions and their partial derivatives, and therefore cannot be solved by commonly used techniques. We are able to solve these PDEs numerically with good accuracy and perform the deconditioning with respect to the queue-length probabilities by evaluating a certain complex integral. Numerical results for the density and the first four moments compare well against regenerative simulation.
加入最短队列(join -the- short - queue, JSQ)是许多网络提供商、云服务器和流量管理系统所选择的调度策略,其中单个队列在处理器共享(processor sharing, PS)队列规则下提供服务。本文首次给出了采用JSQ调度的两个并行PS队列响应时间分布的数值解。利用生成函数方法,得到了与条件响应时间相对应的两个偏微分方程(PDEs),其中条件响应是针对加入第一个或第二个队列的特定跟踪任务。这些偏微分方程是包含偏生成函数及其偏导数的函数方程,因此不能用常用的技术来求解。我们能够以较好的精度对这些偏微分方程进行数值求解,并通过求某个复积分对队列长度概率进行去条件化。密度和前4阶矩的数值模拟结果与再生模拟结果比较吻合。
{"title":"Response time in a pair of processor sharing queues with Join-the-Shortest-Queue scheduling","authors":"Julianna Bor,&nbsp;Peter G. Harrison","doi":"10.1016/j.peva.2025.102509","DOIUrl":"10.1016/j.peva.2025.102509","url":null,"abstract":"<div><div>Join-the-Shortest-Queue (JSQ) is the scheduling policy of choice for many network providers, cloud servers, and traffic management systems, where individual queues are served under the processor sharing (PS) queueing discipline. A numerical solution for the response time distribution in two parallel PS queues with JSQ scheduling is derived for the first time. Using the generating function method, two partial differential equations (PDEs) are obtained corresponding to conditional response times, where the conditioning is on a particular traced task joining the first or the second queue. These PDEs are functional equations that contain partial generating functions and their partial derivatives, and therefore cannot be solved by commonly used techniques. We are able to solve these PDEs numerically with good accuracy and perform the deconditioning with respect to the queue-length probabilities by evaluating a certain complex integral. Numerical results for the density and the first four moments compare well against regenerative simulation.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"170 ","pages":"Article 102509"},"PeriodicalIF":0.8,"publicationDate":"2025-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145099156","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimal strategy against straightforward bidding in clock auctions 时钟拍卖中直接竞价的最优策略
IF 0.8 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-07-23 DOI: 10.1016/j.peva.2025.102502
Jad Zeroual , Marianne Akian , Aurélien Bechler , Matthieu Chardy , Stéphane Gaubert
We study a model of auction representative of the 5G auction in France. We determine the optimal strategy of a bidder, assuming that the valuations of competitors are unknown to this bidder and that competitors adopt the straightforward bidding strategy. Our model is based on a Partially Observable Markov Decision Process (POMDP). This POMDP admits a concise statistics, avoiding the solution of a dynamic programming equation in the space of beliefs. In addition, under this optimal strategy, the expected gain of the bidder does not decrease if competitors deviate from straightforward bidding. We illustrate our results by numerical experiments, comparing the value of the bidder with the value of a perfectly informed one.
我们研究了一个以法国5G拍卖为代表的拍卖模型。我们确定了竞标者的最优策略,假设竞标者不知道竞争对手的估值,且竞争对手采用直接投标策略。我们的模型基于部分可观察马尔可夫决策过程(POMDP)。该POMDP允许简洁的统计,避免了在信念空间求解动态规划方程。此外,在该最优策略下,即使竞争对手偏离直接竞标,竞标者的期望收益也不会降低。我们通过数值实验来说明我们的结果,将出价者的价值与完全知情的价值进行比较。
{"title":"Optimal strategy against straightforward bidding in clock auctions","authors":"Jad Zeroual ,&nbsp;Marianne Akian ,&nbsp;Aurélien Bechler ,&nbsp;Matthieu Chardy ,&nbsp;Stéphane Gaubert","doi":"10.1016/j.peva.2025.102502","DOIUrl":"10.1016/j.peva.2025.102502","url":null,"abstract":"<div><div>We study a model of auction representative of the 5G auction in France. We determine the optimal strategy of a bidder, assuming that the valuations of competitors are unknown to this bidder and that competitors adopt the straightforward bidding strategy. Our model is based on a Partially Observable Markov Decision Process (POMDP). This POMDP admits a concise statistics, avoiding the solution of a dynamic programming equation in the space of beliefs. In addition, under this optimal strategy, the expected gain of the bidder does not decrease if competitors deviate from straightforward bidding. We illustrate our results by numerical experiments, comparing the value of the bidder with the value of a perfectly informed one.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"169 ","pages":"Article 102502"},"PeriodicalIF":0.8,"publicationDate":"2025-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144724472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reliability evaluation of tape library systems 磁带库系统可靠性评估
IF 1 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-06-20 DOI: 10.1016/j.peva.2025.102501
Ilias Iliadis, Mark Lantz
Magnetic tape is a digital data storage technology that has evolved continuously over the last seven decades. It provides a cost-effective way to retain the rapidly increasing volumes of data being created in recent years. The low cost per terabyte combined with tape’s low energy consumption make it an appealing option for storing infrequently accessed data and has resulted in a resurgence in use of the technology. Power and operational failures may damage tapes and lead to data loss. To protect stored data against loss and achieve high data reliability, an erasure coding scheme is employed. A theoretical model capturing the effect of tape failures and latent errors on system reliability is developed. Closed-form expressions are derived for the Mean Time to Data Loss (MTTDL) and the Expected Annual Fraction of Effective Data Loss (EAFEDL) reliability metric, which assesses losses at the file, object, or block, level. The results obtained demonstrate that, for realistic values of bit error rates, reliability is affected by the presence of latent errors. The effect of system parameters on reliability is assessed by conducting a sensitivity evaluation. The reliability improvement achieved by employing erasure coding schemes with increased capability is demonstrated. The theoretical results derived can be used to dimension and provision tape libraries to provide desired levels of data durability.
磁带是一种数字数据存储技术,在过去七十年中不断发展。它提供了一种经济有效的方式来保留近年来创建的快速增长的数据量。每太字节的低成本加上磁带的低能耗,使其成为存储不经常访问的数据的一个有吸引力的选择,并导致了该技术使用的复苏。如果电源或操作异常,可能会导致磁带损坏、数据丢失。为了保护存储的数据不丢失和提高数据的可靠性,采用了erasure编码方案。建立了一个理论模型,描述了磁带故障和潜在错误对系统可靠性的影响。导出了平均数据丢失时间(MTTDL)和有效数据丢失的预期年分数(EAFEDL)可靠性度量的封闭表达式,它们评估文件、对象或块级别的损失。结果表明,对于误码率的实际值,潜在错误的存在会影响可靠性。通过灵敏度评估来评估系统参数对可靠性的影响。证明了采用提高容量的擦除编码方案可以提高系统的可靠性。导出的理论结果可用于定义和提供磁带库,以提供所需的数据持久性水平。
{"title":"Reliability evaluation of tape library systems","authors":"Ilias Iliadis,&nbsp;Mark Lantz","doi":"10.1016/j.peva.2025.102501","DOIUrl":"10.1016/j.peva.2025.102501","url":null,"abstract":"<div><div>Magnetic tape is a digital data storage technology that has evolved continuously over the last seven decades. It provides a cost-effective way to retain the rapidly increasing volumes of data being created in recent years. The low cost per terabyte combined with tape’s low energy consumption make it an appealing option for storing infrequently accessed data and has resulted in a resurgence in use of the technology. Power and operational failures may damage tapes and lead to data loss. To protect stored data against loss and achieve high data reliability, an erasure coding scheme is employed. A theoretical model capturing the effect of tape failures and latent errors on system reliability is developed. Closed-form expressions are derived for the Mean Time to Data Loss (<span><math><mtext>MTTDL</mtext></math></span>) and the Expected Annual Fraction of Effective Data Loss (<span><math><mtext>EAFEDL</mtext></math></span>) reliability metric, which assesses losses at the file, object, or block, level. The results obtained demonstrate that, for realistic values of bit error rates, reliability is affected by the presence of latent errors. The effect of system parameters on reliability is assessed by conducting a sensitivity evaluation. The reliability improvement achieved by employing erasure coding schemes with increased capability is demonstrated. The theoretical results derived can be used to dimension and provision tape libraries to provide desired levels of data durability.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"169 ","pages":"Article 102501"},"PeriodicalIF":1.0,"publicationDate":"2025-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144338331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Virtual caching with apportioned objects for mobile virtual reality 为移动虚拟现实分配对象的虚拟缓存
IF 1 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-06-14 DOI: 10.1016/j.peva.2025.102500
Nader Alfares, George Kesidis
We consider a content-caching system for Virtual or Augmented Reality (VR/AR) that is shared by a number of user groups. The cache could be located in an edge–cloud datacenter and the users could be mobile. Each user group operates its own LRU-list of a certain capacity in the shared cache. The length of objects simultaneously appearing in plural LRU-lists is equally divided (apportioned) among them, i.e., object sharing among the LRUs. We provide a working-set approximation for this system to quickly estimate the cache-hit probabilities for objects apportioned across user groups. We also prove that a solution to the working-set approximation exists. A way to reduce ripple evictions is discussed and some numerical performance results are provided based on Zipf-distributed object popularities. To evaluate our apportioned object sharing system in a VR context, we introduce a benchmark specifically designed for VR streaming, addressing the latency and Quality of Experience (QoE) challenges that arise due to VR’s dynamic fields of view (FoVs) and user synchronization requirements. Using the VR benchmark, we present additional numerical results for cache-hit rates and users’ QoE to illustrate the system’s effectiveness under VR streaming conditions.
我们考虑了一个由多个用户组共享的虚拟或增强现实(VR/AR)的内容缓存系统。缓存可以位于边缘云数据中心,用户可以是移动的。每个用户组在共享缓存中操作自己的一定容量的lru列表。同时出现在多个lru -list中的对象长度在多个lru -list中平均分配,即lru之间的对象共享。我们为该系统提供了一个工作集近似值,以快速估计跨用户组分配的对象的缓存命中概率。我们还证明了工作集近似的一个解的存在。讨论了一种减少波纹驱逐的方法,并给出了基于zipf分布对象流行度的一些数值性能结果。为了在VR环境中评估我们的分配对象共享系统,我们引入了一个专门为VR流设计的基准,解决了由于VR的动态视场(fov)和用户同步要求而产生的延迟和体验质量(QoE)挑战。使用VR基准,我们给出了额外的缓存命中率和用户QoE的数值结果,以说明系统在VR流条件下的有效性。
{"title":"Virtual caching with apportioned objects for mobile virtual reality","authors":"Nader Alfares,&nbsp;George Kesidis","doi":"10.1016/j.peva.2025.102500","DOIUrl":"10.1016/j.peva.2025.102500","url":null,"abstract":"<div><div>We consider a content-caching system for Virtual or Augmented Reality (VR/AR) that is shared by a number of user groups. The cache could be located in an edge–cloud datacenter and the users could be mobile. Each user group operates its own LRU-list of a certain capacity in the shared cache. The length of objects simultaneously appearing in plural LRU-lists is equally divided (apportioned) among them, i.e., object sharing among the LRUs. We provide a working-set approximation for this system to quickly estimate the cache-hit probabilities for objects apportioned across user groups. We also prove that a solution to the working-set approximation exists. A way to reduce ripple evictions is discussed and some numerical performance results are provided based on Zipf-distributed object popularities. To evaluate our apportioned object sharing system in a VR context, we introduce a benchmark specifically designed for VR streaming, addressing the latency and Quality of Experience (QoE) challenges that arise due to VR’s dynamic fields of view (FoVs) and user synchronization requirements. Using the VR benchmark, we present additional numerical results for cache-hit rates and users’ QoE to illustrate the system’s effectiveness under VR streaming conditions.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"169 ","pages":"Article 102500"},"PeriodicalIF":1.0,"publicationDate":"2025-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144314617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Inference for dynamic Erdős–Rényi random graphs under regime switching 状态切换下动态Erdős-Rényi随机图的推理
IF 1 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-06-12 DOI: 10.1016/j.peva.2025.102499
Michel Mandjes , Jiesen Wang
This paper examines a model involving two dynamic Erdős–Rényi random graphs that evolve in parallel, with edges in each graph alternating between being present and absent according to specified on- and off-time distributions. A key feature of our setup is regime switching: the graph that is observed at any given moment depends on the state of an underlying background process, which is modeled as an alternating renewal process. This modeling framework captures a common situation in various real-world applications, where the observed network is influenced by a (typically unobservable) background process. Such scenarios arise, for example, in economics, communication networks, and biological systems.
In our setup we only have access to aggregate quantities such as the number of active edges or the counts of specific subgraphs (such as stars or complete graphs) in the observed graph; importantly, we do not observe the mode. The objective is to estimate the on- and off-time distributions of the edges in each of the two dynamic Erdős–Rényi random graphs, as well as the distribution of time spent in each of the two modes. By employing parametric models for the on- and off-times and the background process, we develop a method of moments approach to estimate the relevant parameters. Experimental evaluations are conducted to demonstrate the effectiveness of the proposed method in recovering these parameters.
本文研究了一个涉及两个并行演化的动态Erdős-Rényi随机图的模型,每个图中的边根据指定的开、关时间分布在存在和不存在之间交替。我们设置的一个关键特征是状态切换:在任何给定时刻观察到的图取决于底层背景进程的状态,该背景进程被建模为交替更新过程。此建模框架捕获了各种实际应用程序中的常见情况,其中观察到的网络受到(通常不可观察的)背景进程的影响。例如,在经济学、通信网络和生物系统中都会出现这种情况。在我们的设置中,我们只能访问聚合量,例如观察图中活动边的数量或特定子图(如星图或完全图)的计数;重要的是,我们没有观察到模式。目标是估计两个动态Erdős-Rényi随机图中每个边的开、关时间分布,以及在每个模式中花费的时间分布。通过采用开关时间和背景过程的参数模型,提出了一种矩量法估计相关参数的方法。实验验证了该方法在恢复这些参数方面的有效性。
{"title":"Inference for dynamic Erdős–Rényi random graphs under regime switching","authors":"Michel Mandjes ,&nbsp;Jiesen Wang","doi":"10.1016/j.peva.2025.102499","DOIUrl":"10.1016/j.peva.2025.102499","url":null,"abstract":"<div><div>This paper examines a model involving two dynamic Erdős–Rényi random graphs that evolve in parallel, with edges in each graph alternating between being present and absent according to specified on- and off-time distributions. A key feature of our setup is regime switching: the graph that is observed at any given moment depends on the state of an underlying background process, which is modeled as an alternating renewal process. This modeling framework captures a common situation in various real-world applications, where the observed network is influenced by a (typically unobservable) background process. Such scenarios arise, for example, in economics, communication networks, and biological systems.</div><div>In our setup we only have access to aggregate quantities such as the number of active edges or the counts of specific subgraphs (such as stars or complete graphs) in the observed graph; importantly, we do not observe the mode. The objective is to estimate the on- and off-time distributions of the edges in each of the two dynamic Erdős–Rényi random graphs, as well as the distribution of time spent in each of the two modes. By employing parametric models for the on- and off-times and the background process, we develop a method of moments approach to estimate the relevant parameters. Experimental evaluations are conducted to demonstrate the effectiveness of the proposed method in recovering these parameters.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"169 ","pages":"Article 102499"},"PeriodicalIF":1.0,"publicationDate":"2025-06-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144314616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Can attacks reduce Age of Information? 攻击能减少信息时代吗?
IF 1 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-06-02 DOI: 10.1016/j.peva.2025.102498
Josu Doncel , Mohamad Assaad
We study a monitoring system in which a single source sends status updates to a monitor through a communication channel. The communication channel is modeled as a queueing system, and we assume that attacks occur following a random process. When an attack occurs, all packets in the queueing system are discarded. While one might expect attacks to always negatively impact system performance, we demonstrate in this paper that, from the perspective of Age of Information (AoI), attacks can in some cases reduce the AoI. Our objective is to identify the conditions under which AoI is reduced and to determine the attack rate that minimizes or reduces AoI. First, we analyze single and tandem M/M/1/1 queues with preemption and show that attacks cannot reduce AoI in these cases. Next, we examine a single M/M/1/1 queue without preemption and establish necessary and sufficient conditions for the existence of an attack rate that minimizes AoI. For this scenario, we also derive an upper bound for the optimal attack rate and prove that it becomes tight when the arrival rate of updates is very high. Through numerical experiments, we observe that attacks can reduce AoI in tandem M/M/1/1 queues without preemption, as well as in preemptive M/M/1/2 and M/M/1/3 queues. Furthermore, we show that the benefit of attacks on AoI increases with the buffer size.
我们研究了一个监控系统,其中单个源通过通信通道向监视器发送状态更新。通信通道被建模为一个排队系统,我们假设攻击发生在一个随机的过程中。当攻击发生时,队列系统中的所有报文将被丢弃。虽然人们可能期望攻击总是对系统性能产生负面影响,但我们在本文中证明,从信息时代(Age of Information, AoI)的角度来看,攻击在某些情况下可以降低AoI。我们的目标是确定减少AoI的条件,并确定最小化或减少AoI的攻击率。首先,我们分析了具有抢占的单队列和串联M/M/1/1队列,并表明在这些情况下攻击不能降低AoI。接下来,我们研究了一个没有抢占的M/M/1/1队列,并建立了使AoI最小的攻击率存在的充分必要条件。对于这种情况,我们还推导了最优攻击率的上界,并证明当更新到达率非常高时,它会变得很紧。通过数值实验,我们观察到攻击可以减少无抢占的M/M/1/1队列中的AoI,以及抢占的M/M/1/2和M/M/1/3队列中的AoI。此外,我们还表明,攻击AoI的好处随着缓冲区大小的增加而增加。
{"title":"Can attacks reduce Age of Information?","authors":"Josu Doncel ,&nbsp;Mohamad Assaad","doi":"10.1016/j.peva.2025.102498","DOIUrl":"10.1016/j.peva.2025.102498","url":null,"abstract":"<div><div>We study a monitoring system in which a single source sends status updates to a monitor through a communication channel. The communication channel is modeled as a queueing system, and we assume that attacks occur following a random process. When an attack occurs, all packets in the queueing system are discarded. While one might expect attacks to always negatively impact system performance, we demonstrate in this paper that, from the perspective of Age of Information (AoI), attacks can in some cases reduce the AoI. Our objective is to identify the conditions under which AoI is reduced and to determine the attack rate that minimizes or reduces AoI. First, we analyze single and tandem M/M/1/1 queues with preemption and show that attacks cannot reduce AoI in these cases. Next, we examine a single M/M/1/1 queue without preemption and establish necessary and sufficient conditions for the existence of an attack rate that minimizes AoI. For this scenario, we also derive an upper bound for the optimal attack rate and prove that it becomes tight when the arrival rate of updates is very high. Through numerical experiments, we observe that attacks can reduce AoI in tandem M/M/1/1 queues without preemption, as well as in preemptive M/M/1/2 and M/M/1/3 queues. Furthermore, we show that the benefit of attacks on AoI increases with the buffer size.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"169 ","pages":"Article 102498"},"PeriodicalIF":1.0,"publicationDate":"2025-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144196352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Performance Evaluation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1