首页 > 最新文献

IEEE Transactions on Cloud Computing最新文献

英文 中文
Cloud-Edge Collaborative Service Architecture With Large-Tiny Models Based on Deep Reinforcement Learning 基于深度强化学习的大-小模型云边缘协同服务架构
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-01-02 DOI: 10.1109/TCC.2024.3525076
Xiaofeng Ji;Faming Gong;Nuanlai Wang;Junjie Xu;Xing Yan
Offshore drilling platforms (ODPs) are critical infrastructure for exploring and developing marine oil and gas resources. As these platforms’ capabilities expand, deploying intelligent surveillance services to ensure safe production has become increasingly important. However, the unique geographical locations and harsh environmental conditions of ODPs pose significant challenges for processing large volumes of video data, complicating the implementation of efficient surveillance systems. This study proposes a Cloud-Edge Large-Tiny Model Collaborative (CELTC) architecture grounded in deep reinforcement learning to optimize the processing and decision-making of surveillance data in offshore drilling platform scenarios. CELTC architecture leverages edge-cloud computing, deploying complex, high-precision large models on cloud servers and lightweight tiny models on edge devices. This dual deployment strategy capitalizes on tiny models’ rapid response and large cloud models’ high-precision capabilities. Additionally, the architecture integrates a deep reinforcement learning algorithm designed to optimize the scheduling and offloading of computational tasks between large and tiny models in the cloud-edge environment. The efficacy of the proposed architecture is validated using real-world surveillance data from ODPs through simulations and comparative experiments.
海洋钻井平台是勘探开发海洋油气资源的重要基础设施。随着这些平台功能的扩展,部署智能监控服务以确保安全生产变得越来越重要。然而,odp独特的地理位置和恶劣的环境条件对处理大量视频数据构成了重大挑战,使高效监控系统的实施复杂化。本研究提出了一种基于深度强化学习的Cloud-Edge Large-Tiny Model Collaborative (CELTC)架构,以优化海上钻井平台场景中监控数据的处理和决策。CELTC架构利用边缘云计算,在云服务器上部署复杂、高精度的大型模型,在边缘设备上部署轻量级的微型模型。这种双重部署策略利用了微型模型的快速响应能力和大型云模型的高精度能力。此外,该架构集成了一种深度强化学习算法,旨在优化云边缘环境中大型和小型模型之间计算任务的调度和卸载。通过模拟和对比实验,利用来自odp的真实世界监控数据验证了所提出架构的有效性。
{"title":"Cloud-Edge Collaborative Service Architecture With Large-Tiny Models Based on Deep Reinforcement Learning","authors":"Xiaofeng Ji;Faming Gong;Nuanlai Wang;Junjie Xu;Xing Yan","doi":"10.1109/TCC.2024.3525076","DOIUrl":"https://doi.org/10.1109/TCC.2024.3525076","url":null,"abstract":"Offshore drilling platforms (ODPs) are critical infrastructure for exploring and developing marine oil and gas resources. As these platforms’ capabilities expand, deploying intelligent surveillance services to ensure safe production has become increasingly important. However, the unique geographical locations and harsh environmental conditions of ODPs pose significant challenges for processing large volumes of video data, complicating the implementation of efficient surveillance systems. This study proposes a Cloud-Edge Large-Tiny Model Collaborative (CELTC) architecture grounded in deep reinforcement learning to optimize the processing and decision-making of surveillance data in offshore drilling platform scenarios. CELTC architecture leverages edge-cloud computing, deploying complex, high-precision large models on cloud servers and lightweight tiny models on edge devices. This dual deployment strategy capitalizes on tiny models’ rapid response and large cloud models’ high-precision capabilities. Additionally, the architecture integrates a deep reinforcement learning algorithm designed to optimize the scheduling and offloading of computational tasks between large and tiny models in the cloud-edge environment. The efficacy of the proposed architecture is validated using real-world surveillance data from ODPs through simulations and comparative experiments.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"288-302"},"PeriodicalIF":5.3,"publicationDate":"2025-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Online Computing Offloading for Budget- Constrained Cloud-Edge Collaborative Video Streaming Systems 预算受限的云边缘协同视频流系统的高效在线计算卸载
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-30 DOI: 10.1109/TCC.2024.3524310
Shijing Yuan;Yuxin Liu;Song Guo;Jie Li;Hongyang Chen;Chentao Wu;Yang Yang
Cloud-Edge Collaborative Architecture (CEA) is a prominent framework that provides low-latency and energy-efficient solutions for video stream processing. In Cloud-Edge Collaborative Video Streaming Systems (CEAVS), efficient online offloading strategies for video tasks are crucial for enhancing user experience. However, most existing works overlook budget constraints, which limits their applicability in real-world scenarios constrained by finite resources. Moreover, they fail to adequately address the heterogeneity of video task redundancies, leading to suboptimal utilization of CEAVS's limited resources. To bridge these gaps, we propose an Efficient Online Computing framework for CEAVS (EOCA) that jointly optimizes accuracy, energy consumption, and latency performance through adaptive online offloading and redundancy compression, without requiring future task information. Technically, we formulate computing offloading and adaptive compression under budget constraints as a stochastic optimization problem that maximizes system satisfaction, defined as a weighted combination of accuracy, latency, and energy performance. We employ Lyapunov optimization to decouple the long-term budget constraint. We prove that the decoupled problem is a generalized ordinal potential game and propose algorithms based on generalized Benders decomposition (GBD) and the best response to obtain Nash equilibrium strategies for computing offloading and task compression. Finally, we analyze EOCA's performance bound, convergence rate, and worst-case performance guarantees. Evaluations demonstrate that EOCA effectively improves satisfaction while effectively balancing satisfaction and computational overhead.
云边缘协作架构(CEA)是一个杰出的框架,为视频流处理提供低延迟和节能的解决方案。在云边缘协同视频流系统(CEAVS)中,高效的视频任务在线卸载策略对于增强用户体验至关重要。然而,大多数现有的作品忽略了预算约束,这限制了它们在有限资源约束下的现实场景中的适用性。此外,它们不能充分解决视频任务冗余的异质性,导致CEAVS有限资源的利用率不理想。为了弥补这些差距,我们提出了一种高效的CEAVS在线计算框架(EOCA),该框架通过自适应在线卸载和冗余压缩共同优化精度,能耗和延迟性能,而不需要未来的任务信息。从技术上讲,我们将预算约束下的计算卸载和自适应压缩表述为最大化系统满意度的随机优化问题,定义为准确性、延迟和能源性能的加权组合。我们使用Lyapunov优化来解耦长期预算约束。我们证明解耦问题是一个广义有序势博弈,并提出了基于广义Benders分解(GBD)和最佳响应的算法来获得计算卸载和任务压缩的纳什均衡策略。最后,我们分析了EOCA的性能边界、收敛速度和最坏情况下的性能保证。评估表明,EOCA有效地提高了满意度,同时有效地平衡了满意度和计算开销。
{"title":"Efficient Online Computing Offloading for Budget- Constrained Cloud-Edge Collaborative Video Streaming Systems","authors":"Shijing Yuan;Yuxin Liu;Song Guo;Jie Li;Hongyang Chen;Chentao Wu;Yang Yang","doi":"10.1109/TCC.2024.3524310","DOIUrl":"https://doi.org/10.1109/TCC.2024.3524310","url":null,"abstract":"Cloud-Edge Collaborative Architecture (CEA) is a prominent framework that provides low-latency and energy-efficient solutions for video stream processing. In Cloud-Edge Collaborative Video Streaming Systems (CEAVS), efficient online offloading strategies for video tasks are crucial for enhancing user experience. However, most existing works overlook budget constraints, which limits their applicability in real-world scenarios constrained by finite resources. Moreover, they fail to adequately address the heterogeneity of video task redundancies, leading to suboptimal utilization of CEAVS's limited resources. To bridge these gaps, we propose an Efficient Online Computing framework for CEAVS (EOCA) that jointly optimizes accuracy, energy consumption, and latency performance through adaptive online offloading and redundancy compression, without requiring future task information. Technically, we formulate computing offloading and adaptive compression under budget constraints as a stochastic optimization problem that maximizes system satisfaction, defined as a weighted combination of accuracy, latency, and energy performance. We employ Lyapunov optimization to decouple the long-term budget constraint. We prove that the decoupled problem is a generalized ordinal potential game and propose algorithms based on generalized Benders decomposition (GBD) and the best response to obtain Nash equilibrium strategies for computing offloading and task compression. Finally, we analyze EOCA's performance bound, convergence rate, and worst-case performance guarantees. Evaluations demonstrate that EOCA effectively improves satisfaction while effectively balancing satisfaction and computational overhead.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"273-287"},"PeriodicalIF":5.3,"publicationDate":"2024-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Differentially Private and Truthful Reverse Auction With Dynamic Resource Provisioning for VNFI Procurement in NFV Markets NFV市场中具有动态资源配置的差异化私有真实反向拍卖
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-26 DOI: 10.1109/TCC.2024.3522963
Xueyi Wang;Xingwei Wang;Zhitong Wang;Rongfei Zeng;Ruiyun Yu;Qiang He;Min Huang
With the advent of network function virtualization (NFV), many users resort to network service provisioning through virtual network function instances (VNFIs) run on the standard physical server in clouds. Following this trend, NFV markets are emerging, which allow a user to procure VNFIs from cloud service providers (CSPs). In such procurement process, it is a significant challenge to ensure differential privacy and truthfulness while explicitly considering dynamic resource provisioning, location sensitiveness and budget of each VNFI. As such, we design a differentially private and truthful reverse auction with dynamic resource provisioning (PTRA-DRP) to resolve the VNFI procurement (VNFIP) problem. To allow dynamic resource provisioning, PTRA-DRP enables CSPs to submit a set of bids and accept as many as possible, and decides the provisioning VNFIs based on the auction outcomes. To be specific, we first devise a greedy heuristic approach to select the set of the winning bids in a differentially privacy-preserving manner. Next, we design a pricing strategy to compute the charges of CSPs, aiming to guarantee truthfulness. Strict theoretical analysis proves that PTRA-DRP can ensure differential privacy, truthfulness, individual rationality, computational efficiency and approximate social cost minimization. Extensive simulations also demonstrate the effectiveness and efficiency of PTRA-DRP.
随着网络功能虚拟化(NFV)的出现,许多用户通过在云中的标准物理服务器上运行的虚拟网络功能实例(VNFIs)来提供网络服务。随着这一趋势,NFV市场正在兴起,允许用户从云服务提供商(csp)那里购买vnfi。在这种采购过程中,在明确考虑每个VNFI的动态资源供应、位置敏感性和预算的同时,确保不同的隐私性和真实性是一个重大挑战。因此,我们设计了一种具有动态资源配置(PTRA-DRP)的不同隐私和真实的反向拍卖来解决VNFI采购(VNFIP)问题。为了实现动态资源供应,PTRA-DRP使csp能够提交一组投标并接受尽可能多的投标,并根据拍卖结果决定供应的vnfi。具体而言,我们首先设计了一种贪婪启发式方法,以差分隐私保护的方式选择中标集。其次,我们设计了一种定价策略来计算csp的费用,以保证其真实性。严格的理论分析证明,PTRA-DRP能够保证差异隐私性、真实性、个体合理性、计算效率和近似的社会成本最小化。大量的仿真也证明了PTRA-DRP算法的有效性和高效性。
{"title":"Differentially Private and Truthful Reverse Auction With Dynamic Resource Provisioning for VNFI Procurement in NFV Markets","authors":"Xueyi Wang;Xingwei Wang;Zhitong Wang;Rongfei Zeng;Ruiyun Yu;Qiang He;Min Huang","doi":"10.1109/TCC.2024.3522963","DOIUrl":"https://doi.org/10.1109/TCC.2024.3522963","url":null,"abstract":"With the advent of network function virtualization (NFV), many users resort to network service provisioning through virtual network function instances (VNFIs) run on the standard physical server in clouds. Following this trend, NFV markets are emerging, which allow a user to procure VNFIs from cloud service providers (CSPs). In such procurement process, it is a significant challenge to ensure differential privacy and truthfulness while explicitly considering dynamic resource provisioning, location sensitiveness and budget of each VNFI. As such, we design a differentially private and truthful reverse auction with dynamic resource provisioning (PTRA-DRP) to resolve the VNFI procurement (VNFIP) problem. To allow dynamic resource provisioning, PTRA-DRP enables CSPs to submit a set of bids and accept as many as possible, and decides the provisioning VNFIs based on the auction outcomes. To be specific, we first devise a greedy heuristic approach to select the set of the winning bids in a differentially privacy-preserving manner. Next, we design a pricing strategy to compute the charges of CSPs, aiming to guarantee truthfulness. Strict theoretical analysis proves that PTRA-DRP can ensure differential privacy, truthfulness, individual rationality, computational efficiency and approximate social cost minimization. Extensive simulations also demonstrate the effectiveness and efficiency of PTRA-DRP.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"259-272"},"PeriodicalIF":5.3,"publicationDate":"2024-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SROdcn: Scalable and Reconfigurable Optical DCN Architecture for High-Performance Computing SROdcn:用于高性能计算的可扩展和可重构光DCN架构
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-26 DOI: 10.1109/TCC.2024.3523433
Kassahun Geresu;Huaxi Gu;Xiaoshan Yu;Meaad Fadhel;Hui Tian;Wenting Wei
Data Center Network (DCN) flexibility is critical for providing adaptive and dynamic bandwidth while optimizing network resources to manage variable traffic patterns generated by heterogeneous applications. To provide flexible bandwidth, this work proposes a machine learning approach with a new Scalable and Reconfigurable Optical DCN (SROdcn) architecture that maintains dynamic and non-uniform network traffic according to the scale of the high-performance optical interconnected DCN. Our main device is the Fiber Optical Switch (FOS), which offers competitive wavelength resolution. We propose a new top-of-rack (ToR) switch that utilizes Wavelength Selective Switches (WSS) to investigate Software-Defined Networking (SDN) with machine learning-enabled flow prediction for reconfigurable optical Data Center Networks (DCNs). Our architecture provides highly scalable and flexible bandwidth allocation. Results from Mininet experimental simulations demonstrate that under the management of an SDN controller, machine learning traffic flow prediction and graph connectivity allow each optical bandwidth to be automatically reconfigured according to variable traffic patterns. The average server-to-server packet delay performance of the reconfigurable SROdcn improves by 42.33% compared to inflexible interconnects. Furthermore, the network performance of flexible SROdcn servers shows up to a 49.67% latency improvement over the Passive Optical Data Center Architecture (PODCA), a 16.87% latency improvement over the optical OPSquare DCN, and up to a 71.13% latency improvement over the fat-tree network. Additionally, our optimized Unsupervised Machine Learning (ML-UnS) method for SROdcn outperforms Supervised Machine Learning (ML-S) and Deep Learning (DL).
数据中心网络(DCN)的灵活性对于提供自适应动态带宽,同时优化网络资源以管理异构应用产生的可变流量模式至关重要。为了提供灵活的带宽,本研究提出了一种机器学习方法,该方法采用了一种新的可扩展和可重构光 DCN(SROdcn)架构,可根据高性能光互连 DCN 的规模保持动态和非均匀的网络流量。我们的主要设备是光纤交换机(FOS),它能提供具有竞争力的波长分辨率。我们提出了一种新的机架顶部(ToR)交换机,利用波长选择开关(WSS)来研究软件定义网络(SDN),并为可重构的光数据中心网络(DCN)提供机器学习功能的流量预测。我们的架构可提供高度可扩展和灵活的带宽分配。Mininet 实验模拟的结果表明,在 SDN 控制器的管理下,机器学习流量预测和图连接允许根据可变流量模式自动重新配置每个光带宽。与不灵活的互联相比,可重新配置的 SROdcn 的服务器到服务器数据包平均延迟性能提高了 42.33%。此外,与无源光数据中心架构(PODCA)相比,灵活的 SROdcn 服务器的网络性能提高了 49.67% 的延迟,与光学 OPSquare DCN 相比提高了 16.87% 的延迟,与胖树网络相比提高了 71.13% 的延迟。此外,我们针对 SROdcn 优化的无监督机器学习(ML-UnS)方法优于有监督机器学习(ML-S)和深度学习(DL)。
{"title":"SROdcn: Scalable and Reconfigurable Optical DCN Architecture for High-Performance Computing","authors":"Kassahun Geresu;Huaxi Gu;Xiaoshan Yu;Meaad Fadhel;Hui Tian;Wenting Wei","doi":"10.1109/TCC.2024.3523433","DOIUrl":"https://doi.org/10.1109/TCC.2024.3523433","url":null,"abstract":"Data Center Network (DCN) flexibility is critical for providing adaptive and dynamic bandwidth while optimizing network resources to manage variable traffic patterns generated by heterogeneous applications. To provide flexible bandwidth, this work proposes a machine learning approach with a new Scalable and Reconfigurable Optical DCN (SROdcn) architecture that maintains dynamic and non-uniform network traffic according to the scale of the high-performance optical interconnected DCN. Our main device is the Fiber Optical Switch (FOS), which offers competitive wavelength resolution. We propose a new top-of-rack (ToR) switch that utilizes Wavelength Selective Switches (WSS) to investigate Software-Defined Networking (SDN) with machine learning-enabled flow prediction for reconfigurable optical Data Center Networks (DCNs). Our architecture provides highly scalable and flexible bandwidth allocation. Results from Mininet experimental simulations demonstrate that under the management of an SDN controller, machine learning traffic flow prediction and graph connectivity allow each optical bandwidth to be automatically reconfigured according to variable traffic patterns. The average server-to-server packet delay performance of the reconfigurable SROdcn improves by 42.33% compared to inflexible interconnects. Furthermore, the network performance of flexible SROdcn servers shows up to a 49.67% latency improvement over the Passive Optical Data Center Architecture (PODCA), a 16.87% latency improvement over the optical OPSquare DCN, and up to a 71.13% latency improvement over the fat-tree network. Additionally, our optimized Unsupervised Machine Learning (ML-UnS) method for SROdcn outperforms Supervised Machine Learning (ML-S) and Deep Learning (DL).","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"245-258"},"PeriodicalIF":5.3,"publicationDate":"2024-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing the Availability and Security of Attestation Scheme for Multiparty-Involved DLaaS: A Circular Approach 提高多参与方DLaaS认证方案的可用性和安全性:一种循环方法
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-26 DOI: 10.1109/TCC.2024.3522993
Miaomiao Yang;Guosheng Huang;Honghai Chen;Yongyi Liao;Qixu Wang;Xingshu Chen
In this paper, we propose a remote attestation approach based on multiple verifiers named CARE. CARE aims to enhance the practicality and efficiency of remote attestation while addressing trust issues within environments involving multiple stakeholders. Specifically, CARE adopts the concept of swarm verification, and employs a circular collaboration model with multiple verifiers to collect and validate evidence, thereby resolving trust issues and enhancing verification efficiency. Moreover, CARE introduces a meticulously designed filtering mechanism to address the issue of false positives in verification outcomes non-invasively. CARE utilizes a multiway tree structure to construct the baseline value library, which enhances the flexibility and fine-grained management capability of the system. Security analysis indicates that CARE can effectively resist collusion attacks. Further, detailed simulation experiments have validated its capability to convincingly attest to the trustworthiness of the dynamically constructed environment. Notably, CARE is also suitable for the remote attestation of large-scale virtual machines, achieving an efficiency 9 times greater than the classical practice approach. To the best of our knowledge, CARE is the first practical solution to address inaccuracies in remote attestation results caused by the activation of Integrity Measurement Architecture (IMA) at the application layer.
在本文中,我们提出了一种基于多个验证者的远程认证方法——CARE。CARE旨在提高远程认证的实用性和效率,同时解决涉及多个利益相关者的环境中的信任问题。具体而言,CARE采用群体验证的概念,采用多个验证者的循环协作模型来收集和验证证据,从而解决信任问题,提高验证效率。此外,CARE引入了精心设计的过滤机制,以非侵入性地解决验证结果中的误报问题。CARE采用多路树结构构建基线值库,增强了系统的灵活性和细粒度管理能力。安全分析表明,CARE能够有效抵御合谋攻击。此外,详细的仿真实验验证了该方法能够令人信服地证明动态构建环境的可信度。值得注意的是,CARE也适用于大型虚拟机的远程认证,其效率是传统实践方法的9倍。据我们所知,CARE是解决由于在应用层激活完整性度量体系结构(IMA)而导致的远程认证结果不准确的第一个实用解决方案。
{"title":"Enhancing the Availability and Security of Attestation Scheme for Multiparty-Involved DLaaS: A Circular Approach","authors":"Miaomiao Yang;Guosheng Huang;Honghai Chen;Yongyi Liao;Qixu Wang;Xingshu Chen","doi":"10.1109/TCC.2024.3522993","DOIUrl":"https://doi.org/10.1109/TCC.2024.3522993","url":null,"abstract":"In this paper, we propose a remote attestation approach based on multiple verifiers named CARE. CARE aims to enhance the practicality and efficiency of remote attestation while addressing trust issues within environments involving multiple stakeholders. Specifically, CARE adopts the concept of swarm verification, and employs a circular collaboration model with multiple verifiers to collect and validate evidence, thereby resolving trust issues and enhancing verification efficiency. Moreover, CARE introduces a meticulously designed filtering mechanism to address the issue of false positives in verification outcomes non-invasively. CARE utilizes a multiway tree structure to construct the baseline value library, which enhances the flexibility and fine-grained management capability of the system. Security analysis indicates that CARE can effectively resist collusion attacks. Further, detailed simulation experiments have validated its capability to convincingly attest to the trustworthiness of the dynamically constructed environment. Notably, CARE is also suitable for the remote attestation of large-scale virtual machines, achieving an efficiency 9 times greater than the classical practice approach. To the best of our knowledge, CARE is the first practical solution to address inaccuracies in remote attestation results caused by the activation of Integrity Measurement Architecture (IMA) at the application layer.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"227-244"},"PeriodicalIF":5.3,"publicationDate":"2024-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570598","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing Sustainability in Data Centers: Evaluation of Hybrid Air/Liquid Cooling Schemes for IT Payload Using Sea Water 推进数据中心的可持续性:评估使用海水的IT有效载荷的混合空气/液体冷却方案
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-24 DOI: 10.1109/TCC.2024.3521666
Imran Latif;Muhammad Mubashar Ashraf;Umaima Haider;Gemma Reeves;Alexandrina Untaroiu;Fábio Coelho;Denis Browne
The growth in cloud computing, Big Data, AI and high-performance computing (HPC) necessitate the deployment of additional data centers (DC’s) with high energy demands. The unprecedented increase in the Thermal Design Power (TDP) of the computing chips will require innovative cooling techniques. Furthermore, DC’s are increasingly limited in their ability to add powerful GPU servers by power capacity constraints. As cooling energy use accounts for up to 40% of DC energy consumption, creative cooling solutions are urgently needed to allow deployment of additional servers, enhance sustainability and increase energy efficiency of DC’s. The information in this study is provided from Start Campus’ Sines facility supported by Alfa Laval for the heat exchanger and CO2 emission calculations. The study evaluates the performance and sustainability impact of various data center cooling strategies including an air-only deployment and a subsequent hybrid air/water cooling solution all utilizing sea water as the cooling source. We evaluate scenarios from 3 MW to 15+1 MW of IT load in 3 MW increments which correspond to the size of heat exchangers used in the Start Campus’ modular system design. This study also evaluates the CO2 emissions compared to a conventional chiller system for all the presented scenarios. Results indicate that the effective use of the sea water cooled system combined with liquid cooled systems improve the efficiency of the DC, plays a role in decreasing the CO2 emissions and supports in achieving sustainability goals.
云计算、大数据、人工智能和高性能计算(HPC)的增长需要部署额外的高能耗数据中心(DC)。计算芯片的热设计功率(TDP)的空前增长将需要创新的冷却技术。此外,由于功率容量的限制,数据中心在添加强大的GPU服务器方面的能力越来越有限。由于冷却能耗占直流能耗的40%,因此迫切需要创造性的冷却解决方案,以允许部署额外的服务器,增强可持续性并提高直流的能源效率。本研究中的信息由阿法拉伐公司支持的Start Campus的Sines设施提供,用于换热器和二氧化碳排放计算。该研究评估了各种数据中心冷却策略的性能和可持续性影响,包括纯空气部署和随后使用海水作为冷却源的混合空气/水冷却解决方案。我们评估了从3兆瓦到15+1兆瓦IT负载的场景,增量为3兆瓦,对应于Start Campus模块化系统设计中使用的热交换器的大小。本研究还评估了二氧化碳排放量相比,传统的冷水机组系统的所有提出的方案。结果表明,海水冷却系统与液冷系统的有效结合提高了直流系统的效率,减少了二氧化碳的排放,有助于实现可持续发展目标。
{"title":"Advancing Sustainability in Data Centers: Evaluation of Hybrid Air/Liquid Cooling Schemes for IT Payload Using Sea Water","authors":"Imran Latif;Muhammad Mubashar Ashraf;Umaima Haider;Gemma Reeves;Alexandrina Untaroiu;Fábio Coelho;Denis Browne","doi":"10.1109/TCC.2024.3521666","DOIUrl":"https://doi.org/10.1109/TCC.2024.3521666","url":null,"abstract":"The growth in cloud computing, Big Data, AI and high-performance computing (HPC) necessitate the deployment of additional data centers (DC’s) with high energy demands. The unprecedented increase in the Thermal Design Power (TDP) of the computing chips will require innovative cooling techniques. Furthermore, DC’s are increasingly limited in their ability to add powerful GPU servers by power capacity constraints. As cooling energy use accounts for up to 40% of DC energy consumption, creative cooling solutions are urgently needed to allow deployment of additional servers, enhance sustainability and increase energy efficiency of DC’s. The information in this study is provided from Start Campus’ Sines facility supported by Alfa Laval for the heat exchanger and CO<sub>2</sub> emission calculations. The study evaluates the performance and sustainability impact of various data center cooling strategies including an air-only deployment and a subsequent hybrid air/water cooling solution all utilizing sea water as the cooling source. We evaluate scenarios from 3 MW to 15+1 MW of IT load in 3 MW increments which correspond to the size of heat exchangers used in the Start Campus’ modular system design. This study also evaluates the CO<sub>2</sub> emissions compared to a conventional chiller system for all the presented scenarios. Results indicate that the effective use of the sea water cooled system combined with liquid cooled systems improve the efficiency of the DC, plays a role in decreasing the CO<sub>2</sub> emissions and supports in achieving sustainability goals.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"184-197"},"PeriodicalIF":5.3,"publicationDate":"2024-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding Serverless Inference in Mobile-Edge Networks: A Benchmark Approach 理解移动边缘网络中的无服务器推理:基准方法
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-24 DOI: 10.1109/TCC.2024.3521657
Junhong Chen;Yanying Lin;Shijie Peng;Shuaipeng Wu;Kenneth Kent;Hao Dai;Kejiang Ye;Yang Wang
Although the emerging serverless paradigm has the potential to become a dominant way of deploying cloud-service tasks across millions of mobile and IoT devices, the overhead characteristics of executing these tasks on such a volume of mobile devices remain largely unclear. To address this issue, this paper conducts a deep analysis based on the OpenFaaS platform—a popular open-source serverless platform for mobile edge environments—to investigate the overhead of performing deep learning inference tasks on mobile devices. To thoroughly evaluate the inference overhead, we develop a performance benchmark, named ESBench, whereby a set of comprehensive experiments are conducted with respect to a bunch of simulated mobile devices associated with an edge cluster. Our investigation reveals that the performance of deep learning inference tasks is significantly influenced by the model size and resource contention in mobile devices, leading to up to $3times$ degradation in performance. Moreover, we observe that the network environment can negatively impact the performance of mobile inference, increasing the CPU overhead under poor network conditions. Based on our findings, we further propose some recommendations for designing efficient serverless platforms and resource management strategies as well as for deploying serverless computing in the mobile edge environment.
尽管新兴的无服务器范式有可能成为在数百万移动和物联网设备上部署云服务任务的主要方式,但在如此大量的移动设备上执行这些任务的开销特征在很大程度上仍不清楚。为了解决这个问题,本文基于OpenFaaS平台(一种流行的用于移动边缘环境的开源无服务器平台)进行了深入分析,以调查在移动设备上执行深度学习推理任务的开销。为了彻底评估推理开销,我们开发了一个名为ESBench的性能基准,其中针对与边缘集群相关的一堆模拟移动设备进行了一组全面的实验。我们的研究表明,深度学习推理任务的性能受到移动设备中模型大小和资源争用的显著影响,导致性能下降高达3倍。此外,我们观察到网络环境会对移动推理的性能产生负面影响,在恶劣的网络条件下会增加CPU开销。基于我们的研究结果,我们进一步提出了一些建议,用于设计高效的无服务器平台和资源管理策略,以及在移动边缘环境中部署无服务器计算。
{"title":"Understanding Serverless Inference in Mobile-Edge Networks: A Benchmark Approach","authors":"Junhong Chen;Yanying Lin;Shijie Peng;Shuaipeng Wu;Kenneth Kent;Hao Dai;Kejiang Ye;Yang Wang","doi":"10.1109/TCC.2024.3521657","DOIUrl":"https://doi.org/10.1109/TCC.2024.3521657","url":null,"abstract":"Although the emerging serverless paradigm has the potential to become a dominant way of deploying cloud-service tasks across millions of mobile and IoT devices, the overhead characteristics of executing these tasks on such a volume of mobile devices remain largely unclear. To address this issue, this paper conducts a deep analysis based on the OpenFaaS platform—a popular open-source serverless platform for mobile edge environments—to investigate the overhead of performing deep learning inference tasks on mobile devices. To thoroughly evaluate the inference overhead, we develop a performance benchmark, named <i>ESBench</i>, whereby a set of comprehensive experiments are conducted with respect to a bunch of simulated mobile devices associated with an edge cluster. Our investigation reveals that the performance of deep learning inference tasks is significantly influenced by the model size and resource contention in mobile devices, leading to up to <inline-formula><tex-math>$3times$</tex-math></inline-formula> degradation in performance. Moreover, we observe that the network environment can negatively impact the performance of mobile inference, increasing the CPU overhead under poor network conditions. Based on our findings, we further propose some recommendations for designing efficient serverless platforms and resource management strategies as well as for deploying serverless computing in the mobile edge environment.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"198-212"},"PeriodicalIF":5.3,"publicationDate":"2024-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
StreamSys: A Lightweight Executable Delivery System for Edge Computing StreamSys:用于边缘计算的轻量级可执行文件交付系统
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-24 DOI: 10.1109/TCC.2024.3521978
Jun Lu;Zhenya Ma;Yinggang Gao;Sheng Yue;Ju Ren;Yaoxue Zhang
Edge computing brings several challenges when it comes to data movement. First, moving large data from edge devices to the server is likely to waste bandwidth. Second, complex data patterns (e.g., traffic cameras) on devices require flexible handling. An ideal approach is to move code to data instead. However, since only a small portion of code is required, moving the executable as well as their libraries to the devices can be an overkill. While loading code on demand from remote such as NFS can be a stopgap, but on the other hand leads to low efficiency for irregular access patterns. This article presents StreamSys, a lightweight executable delivery system that loads code on demand by redirecting the local disk IO to the server through optimized network IO. We employ a Markov-based prefetch mechanism on the server side. It learns the access pattern of code and predicts the block sequence for the client to reduce the network round trip. Meanwhile, server-side StreamSys asynchronously prereads the block sequence from the disk to conceal disk IO latency beforehand. Evaluation shows that the latency of StreamSys is up to 71.4% lower than the native Linux file system based on SD card and up to 62% lower than NFS in wired environments.
当涉及到数据移动时,边缘计算带来了几个挑战。首先,将大数据从边缘设备移动到服务器可能会浪费带宽。其次,设备上复杂的数据模式(如交通摄像头)需要灵活处理。一种理想的方法是将代码移到数据中。然而,由于只需要一小部分代码,因此将可执行文件及其库移动到设备上可能是多余的。虽然从远程(如NFS)按需加载代码可能是权宜之计,但另一方面,对于不规则的访问模式会导致效率低下。本文介绍了StreamSys,这是一个轻量级的可执行交付系统,它通过优化的网络IO将本地磁盘IO重定向到服务器,按需加载代码。我们在服务器端采用了基于markov的预取机制。它学习代码的访问模式,并为客户端预测块序列,以减少网络往返。同时,服务器端StreamSys异步地从磁盘预读块序列,以预先隐藏磁盘IO延迟。评估表明,StreamSys的延迟比基于SD卡的本地Linux文件系统低71.4%,比有线环境下的NFS低62%。
{"title":"StreamSys: A Lightweight Executable Delivery System for Edge Computing","authors":"Jun Lu;Zhenya Ma;Yinggang Gao;Sheng Yue;Ju Ren;Yaoxue Zhang","doi":"10.1109/TCC.2024.3521978","DOIUrl":"https://doi.org/10.1109/TCC.2024.3521978","url":null,"abstract":"Edge computing brings several challenges when it comes to data movement. First, moving large data from edge devices to the server is likely to waste bandwidth. Second, complex data patterns (e.g., traffic cameras) on devices require flexible handling. An ideal approach is to move code to data instead. However, since only a small portion of code is required, moving the executable as well as their libraries to the devices can be an overkill. While loading code on demand from remote such as NFS can be a stopgap, but on the other hand leads to low efficiency for irregular access patterns. This article presents <sc>StreamSys</small>, a lightweight executable delivery system that loads code on demand by redirecting the local disk IO to the server through optimized network IO. We employ a Markov-based prefetch mechanism on the server side. It learns the access pattern of code and predicts the block sequence for the client to reduce the network round trip. Meanwhile, server-side <sc>StreamSys</small> asynchronously prereads the block sequence from the disk to conceal disk IO latency beforehand. Evaluation shows that the latency of <sc>StreamSys</small> is up to 71.4% lower than the native Linux file system based on SD card and up to 62% lower than NFS in wired environments.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"213-226"},"PeriodicalIF":5.3,"publicationDate":"2024-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI Applications Resource Allocation in Computing Continuum: A Stackelberg Game Approach 计算连续体中的人工智能应用资源分配:一个Stackelberg博弈方法
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-20 DOI: 10.1109/TCC.2024.3521213
Roberto Sala;Hamta Sedghani;Mauro Passacantando;Giacomo Verticale;Danilo Ardagna
The growth, development, and commercialization of artificial intelligence-based technologies such as self-driving cars, augmented-reality viewers, chatbots, and virtual assistants are driving the need for increased computing power. Most of these applications rely on Deep Neural Networks (DNNs), which demand substantial computing capacity to meet user demands. However, this capacity cannot be fully provided by users’ local devices due to their limited processing power, nor by cloud data centers due to high transmission latency from long distances. Edge cloud computing addresses this issue by processing user requests through 5G, which reduces transmission latency from local devices to computing resources and allows the offloading of some computations to cloud back-ends. This paper introduces a model for a Mobile Edge Cloud system designed for an application based on a DNN. The interaction among multiple mobile users and the edge platform is formulated as a one-leader multi-follower Stackelberg game, resulting in a challenging non-convex mixed integer nonlinear programming (MINLP) problem. To tackle this, we propose a heuristic approach based on Karush-Kuhn-Tucker conditions, which solves the MINLP problem significantly faster than the commercial state-of-the-art solvers (up to 50,000 times). Furthermore, we present an algorithm to estimate optimal platform profit when sensitive user parameters are unknown. Comparing this with the full-knowledge scenario, we observe a profit loss of approximately 1%. Lastly, we analyze the advantages for an edge provider to engage in a Stackelberg game rather than setting a fixed price for its users, showing potential profit increases ranging from 16% to 66%.
自动驾驶汽车、增强现实观看器、聊天机器人和虚拟助手等基于人工智能的技术的增长、发展和商业化正在推动对增强计算能力的需求。这些应用大多依赖于深度神经网络(dnn),这需要大量的计算能力来满足用户的需求。但是,由于用户本地设备的处理能力有限,无法完全提供这种容量,而云数据中心由于远距离传输的高延迟,也无法完全提供这种容量。边缘云计算通过5G处理用户请求来解决这个问题,这减少了从本地设备到计算资源的传输延迟,并允许将一些计算卸载到云后端。本文介绍了一种基于深度神经网络的移动边缘云系统模型。将多个移动用户与边缘平台之间的交互作用表述为一个领头多follower的Stackelberg博弈,从而产生一个具有挑战性的非凸混合整数非线性规划(MINLP)问题。为了解决这个问题,我们提出了一种基于Karush-Kuhn-Tucker条件的启发式方法,它解决MINLP问题的速度明显快于商业最先进的解决方案(高达50,000次)。此外,我们还提出了一种在敏感用户参数未知的情况下估计最优平台利润的算法。与完全了解情况相比,我们观察到利润损失约为1%。最后,我们分析了边缘提供商参与Stackelberg游戏而不是为其用户设定固定价格的优势,显示潜在利润增长从16%到66%不等。
{"title":"AI Applications Resource Allocation in Computing Continuum: A Stackelberg Game Approach","authors":"Roberto Sala;Hamta Sedghani;Mauro Passacantando;Giacomo Verticale;Danilo Ardagna","doi":"10.1109/TCC.2024.3521213","DOIUrl":"https://doi.org/10.1109/TCC.2024.3521213","url":null,"abstract":"The growth, development, and commercialization of artificial intelligence-based technologies such as self-driving cars, augmented-reality viewers, chatbots, and virtual assistants are driving the need for increased computing power. Most of these applications rely on Deep Neural Networks (DNNs), which demand substantial computing capacity to meet user demands. However, this capacity cannot be fully provided by users’ local devices due to their limited processing power, nor by cloud data centers due to high transmission latency from long distances. Edge cloud computing addresses this issue by processing user requests through 5G, which reduces transmission latency from local devices to computing resources and allows the offloading of some computations to cloud back-ends. This paper introduces a model for a Mobile Edge Cloud system designed for an application based on a DNN. The interaction among multiple mobile users and the edge platform is formulated as a one-leader multi-follower Stackelberg game, resulting in a challenging non-convex mixed integer nonlinear programming (MINLP) problem. To tackle this, we propose a heuristic approach based on Karush-Kuhn-Tucker conditions, which solves the MINLP problem significantly faster than the commercial state-of-the-art solvers (up to 50,000 times). Furthermore, we present an algorithm to estimate optimal platform profit when sensitive user parameters are unknown. Comparing this with the full-knowledge scenario, we observe a profit loss of approximately 1%. Lastly, we analyze the advantages for an edge provider to engage in a Stackelberg game rather than setting a fixed price for its users, showing potential profit increases ranging from 16% to 66%.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"166-183"},"PeriodicalIF":5.3,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Two-Stage Learning Approach for Semantic-Aware Task Scheduling in Container-Based Clouds 基于容器云中语义感知任务调度的两阶段学习方法
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-19 DOI: 10.1109/TCC.2024.3520101
Lilu Zhu;Kai Huang;Yanfeng Hu;Yang Wang
Container-based task scheduling is critical for ensuring a reliable, flexible and cost-effective cloud computing mode. However, in different business cloud systems, state-of-the-art scheduling models are not as effective as those in the simulated world due to the sparsity issues associated with sample sizes and features. Herein, we propose a novel containerized task scheduling framework (SA2CTS) based on reinforcement learning (RL) that incorporates cross-modal contrastive learning (CL) loss. This framework optimizes the scheduler's understanding of the container-based cloud state in RL by adding a pretraining stage, promoting accurate scheduling action inference. Specifically, we design a two-stage learning pipeline. The initial stage involves pretraining the model on a large collection of aligned image-text pairs to extract fine-grained scheduling affinity features, and the high-level semantic representations of scheduling tasks are learned in the multimodal space. In the second stage, we fine-tune the pretrained model with multisource cluster feedback, i.e., build a mapping from state representations to scheduling actions through the RL paradigm, achieving task-oriented and semantic-aware scheduling. The experimental results obtained on three large-scale production cluster datasets substantiate that the proposed SA2CTS method can provide average convergence efficiency and resource utilization improvements of 17.57% and 10.42%, respectively, over the state-of-the-art RL scheduling methods.
基于容器的任务调度对于确保可靠、灵活和经济高效的云计算模式至关重要。然而,在不同的业务云系统中,由于与样本大小和特征相关的稀疏性问题,最先进的调度模型不如模拟世界中的调度模型有效。在此,我们提出了一种新的基于强化学习(RL)的容器化任务调度框架(SA2CTS),该框架结合了跨模态对比学习(CL)损失。该框架通过增加预训练阶段,优化了调度程序对RL中基于容器的云状态的理解,促进了准确的调度动作推断。具体来说,我们设计了一个两阶段的学习管道。初始阶段包括在大量对齐的图像-文本对上对模型进行预训练,以提取细粒度的调度关联特征,并在多模态空间中学习调度任务的高级语义表示。在第二阶段,我们利用多源聚类反馈对预训练模型进行微调,即通过RL范式构建从状态表示到调度动作的映射,实现面向任务和语义感知的调度。在3个大规模生产集群数据集上的实验结果表明,与现有的RL调度方法相比,该方法的平均收敛效率和资源利用率分别提高了17.57%和10.42%。
{"title":"Two-Stage Learning Approach for Semantic-Aware Task Scheduling in Container-Based Clouds","authors":"Lilu Zhu;Kai Huang;Yanfeng Hu;Yang Wang","doi":"10.1109/TCC.2024.3520101","DOIUrl":"https://doi.org/10.1109/TCC.2024.3520101","url":null,"abstract":"Container-based task scheduling is critical for ensuring a reliable, flexible and cost-effective cloud computing mode. However, in different business cloud systems, state-of-the-art scheduling models are not as effective as those in the simulated world due to the sparsity issues associated with sample sizes and features. Herein, we propose a novel containerized task scheduling framework (SA2CTS) based on reinforcement learning (RL) that incorporates cross-modal contrastive learning (CL) loss. This framework optimizes the scheduler's understanding of the container-based cloud state in RL by adding a pretraining stage, promoting accurate scheduling action inference. Specifically, we design a two-stage learning pipeline. The initial stage involves pretraining the model on a large collection of aligned image-text pairs to extract fine-grained scheduling affinity features, and the high-level semantic representations of scheduling tasks are learned in the multimodal space. In the second stage, we fine-tune the pretrained model with multisource cluster feedback, i.e., build a mapping from state representations to scheduling actions through the RL paradigm, achieving task-oriented and semantic-aware scheduling. The experimental results obtained on three large-scale production cluster datasets substantiate that the proposed SA2CTS method can provide average convergence efficiency and resource utilization improvements of 17.57% and 10.42%, respectively, over the state-of-the-art RL scheduling methods.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 1","pages":"148-165"},"PeriodicalIF":5.3,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Cloud Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1