首页 > 最新文献

IEEE Transactions on Cloud Computing最新文献

英文 中文
Accelerating AI-Generated Content Collaborative Inference Via Transfer Reinforcement Learning in Dynamic Edge Networks 在动态边缘网络中通过迁移强化学习加速人工智能生成内容的协同推理
IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-07-10 DOI: 10.1109/TCC.2025.3586878
Meng Tian;Zhicheng Liu;Chenxuan Hou;Chao Qiu;Xiaofei Wang;Dusit Niyato;Victor C. M. Leung
While diffusion models have demonstrated remarkable success in computer vision tasks, their deployment in Internet of Things environments remains challenging. Edge devices face significant constraints in computational resources and must adapt to dynamic operating conditions. To address these limitations, we propose a novel system that accelerates AI-generated content (AIGC) collaborative inference in dynamic edge networks. The proposed system introduces a multi-exit vision transformer-based U-Net architecture that enables efficient processing through adaptive exit point selection during the diffusion process, optimizing the trade-off between inference accuracy and computational efficiency. To optimize device-level operations, we develop an innovative generative AI-assisted reinforcement learning framework that determines optimal exit selection and offloading strategies to maximize generation quality and inference speed. Furthermore, we design a fine-tuning approach with policy reuse mechanisms that facilitates rapid reinforcement learning algorithm deployment across diverse environments. Extensive experimental evaluations demonstrate that our system outperforms existing algorithms in terms of balancing inference latency and generation quality, while also exhibiting improved adaptability to environmental variations.
虽然扩散模型在计算机视觉任务中取得了显著的成功,但在物联网环境中的部署仍然具有挑战性。边缘设备在计算资源方面面临重大限制,必须适应动态操作条件。为了解决这些限制,我们提出了一种在动态边缘网络中加速人工智能生成内容(AIGC)协同推理的新系统。该系统引入了基于多出口视觉转换器的U-Net架构,通过在扩散过程中自适应选择出口点来实现高效处理,优化了推理精度和计算效率之间的权衡。为了优化设备级操作,我们开发了一种创新的生成式人工智能辅助强化学习框架,该框架可确定最佳退出选择和卸载策略,以最大限度地提高生成质量和推理速度。此外,我们设计了一种带有策略重用机制的微调方法,有助于在不同环境中快速部署强化学习算法。大量的实验评估表明,我们的系统在平衡推理延迟和生成质量方面优于现有算法,同时也表现出对环境变化的更好的适应性。
{"title":"Accelerating AI-Generated Content Collaborative Inference Via Transfer Reinforcement Learning in Dynamic Edge Networks","authors":"Meng Tian;Zhicheng Liu;Chenxuan Hou;Chao Qiu;Xiaofei Wang;Dusit Niyato;Victor C. M. Leung","doi":"10.1109/TCC.2025.3586878","DOIUrl":"https://doi.org/10.1109/TCC.2025.3586878","url":null,"abstract":"While diffusion models have demonstrated remarkable success in computer vision tasks, their deployment in Internet of Things environments remains challenging. Edge devices face significant constraints in computational resources and must adapt to dynamic operating conditions. To address these limitations, we propose a novel system that accelerates AI-generated content (AIGC) collaborative inference in dynamic edge networks. The proposed system introduces a multi-exit vision transformer-based U-Net architecture that enables efficient processing through adaptive exit point selection during the diffusion process, optimizing the trade-off between inference accuracy and computational efficiency. To optimize device-level operations, we develop an innovative generative AI-assisted reinforcement learning framework that determines optimal exit selection and offloading strategies to maximize generation quality and inference speed. Furthermore, we design a fine-tuning approach with policy reuse mechanisms that facilitates rapid reinforcement learning algorithm deployment across diverse environments. Extensive experimental evaluations demonstrate that our system outperforms existing algorithms in terms of balancing inference latency and generation quality, while also exhibiting improved adaptability to environmental variations.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 3","pages":"1011-1025"},"PeriodicalIF":5.0,"publicationDate":"2025-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144997132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fog-Enhanced Personalized Privacy-Preserving Data Analysis for Smart Homes 智能家居的雾增强个性化隐私保护数据分析
IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-07-04 DOI: 10.1109/TCC.2025.3586052
Jiajun Chen;Chunqiang Hu;Weihong Sheng;Hui Xia;Pengfei Hu;Jiguo Yu
The proliferation of Internet of Things (IoT) devices has led to a surge in data generation within smart home environments. This data explosion has raised significant privacy concerns and highlighted a lack of user-friendly controls. Consequently, there is a pressing need for a robust privacy-enhancing mechanism tailored for smart homes, safeguarding sensitive data from a user-centric perspective. In this article, we introduce the Fog-enhanced Personalized Differential Privacy (FEPDP) model, which utilizes the distributed nature of fog computing to improve data processing efficiency and security in smart homes. Specifically, the personalization, as a key feature of FEPDP, is manifested through an array of user-driven policy specifications, enabling home users to specify secret and privacy specifications for their personal data. These specifications not only enhance control over personal data but also align with the heterogeneous nature of smart home environments. Subsequently, aligned with fog-based smart home architecture, we propose two policy-driven partitioning mechanisms that utilize threshold partitioning based on dynamic programming to effectively implement FEPDP. Finally, comprehensive theoretical analysis and experimental validation across various statistical analysis tasks and datasets confirm that FEPDP achieves a superior privacy-utility trade-off for smart home data by leveraging non-sensitive data and fog-based partitioning.
物联网(IoT)设备的激增导致智能家居环境中数据生成的激增。这种数据爆炸引起了人们对隐私的严重担忧,并凸显了用户友好控制的缺乏。因此,迫切需要一种为智能家居量身定制的强大的隐私增强机制,从以用户为中心的角度保护敏感数据。在本文中,我们介绍了雾增强的个性化差异隐私(FEPDP)模型,该模型利用雾计算的分布式特性来提高智能家居中的数据处理效率和安全性。具体来说,个性化作为FEPDP的一个关键特征,通过一系列用户驱动的策略规范来体现,使家庭用户能够为其个人数据指定秘密和隐私规范。这些规范不仅增强了对个人数据的控制,而且与智能家居环境的异构特性保持一致。随后,结合基于雾的智能家居架构,我们提出了两种策略驱动的分区机制,利用基于动态规划的阈值分区来有效实现FEPDP。最后,对各种统计分析任务和数据集进行了全面的理论分析和实验验证,证实了FEPDP通过利用非敏感数据和基于雾的分区,实现了智能家居数据的卓越隐私效用权衡。
{"title":"Fog-Enhanced Personalized Privacy-Preserving Data Analysis for Smart Homes","authors":"Jiajun Chen;Chunqiang Hu;Weihong Sheng;Hui Xia;Pengfei Hu;Jiguo Yu","doi":"10.1109/TCC.2025.3586052","DOIUrl":"https://doi.org/10.1109/TCC.2025.3586052","url":null,"abstract":"The proliferation of Internet of Things (IoT) devices has led to a surge in data generation within smart home environments. This data explosion has raised significant privacy concerns and highlighted a lack of user-friendly controls. Consequently, there is a pressing need for a robust privacy-enhancing mechanism tailored for smart homes, safeguarding sensitive data from a user-centric perspective. In this article, we introduce the Fog-enhanced Personalized Differential Privacy (FEPDP) model, which utilizes the distributed nature of fog computing to improve data processing efficiency and security in smart homes. Specifically, the personalization, as a key feature of FEPDP, is manifested through an array of user-driven policy specifications, enabling home users to specify secret and privacy specifications for their personal data. These specifications not only enhance control over personal data but also align with the heterogeneous nature of smart home environments. Subsequently, aligned with fog-based smart home architecture, we propose two policy-driven partitioning mechanisms that utilize threshold partitioning based on dynamic programming to effectively implement FEPDP. Finally, comprehensive theoretical analysis and experimental validation across various statistical analysis tasks and datasets confirm that FEPDP achieves a superior privacy-utility trade-off for smart home data by leveraging non-sensitive data and fog-based partitioning.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 3","pages":"995-1010"},"PeriodicalIF":5.0,"publicationDate":"2025-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144998331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Refrain From Inquiring About My Scalable Storage and Boolean Queries for Secure Cloud 不要询问我的可扩展存储和布尔查询安全云
IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-06-24 DOI: 10.1109/TCC.2025.3582645
Boli Hu;Kai Zhang;Junqing Gong;Haifeng Qian
Outsourcing personal data to a convenient and affordable cloud platform has become a popular practice. Considering the risk of privacy leakage, users usually encrypt their data before uploading it to the cloud server. Searchable encryption (SE) allows cloud servers to manage and search data in encrypted form based on user-specified requests. However, coercion attacks are rarely considered, where users may be forced to open search records and results. Therefore, deniable SE solutions against coercion attacks are presented, but they suffer from large storage overhead or fail to consider the dual coercion situation towards both sides of data owners and data users. In this paper, we roughly combine oblivious cross-tags protocol (OXT) and deniable encryption to propose a deniable SE (deniable cross-tag, DXT) scheme, which supports boolean queries and resists dual coercion attacks. Technically, we formalize a new primitive called updatable deniable encryption, and combine it with OXT in a non-trivial manner. In addition, we give formal system model, security model, and security proof of DXT. By employing the HUAWEI cloud platform, we conduct sufficient comparative experiments between DXT and state-of-the-art solutions based on a public dataset. The experimental results demonstrate that DXT outperforms higher search efficiency while achieving better features.
将个人数据外包给一个方便且价格合理的云平台已经成为一种流行的做法。考虑到隐私泄露的风险,用户在将数据上传到云服务器之前通常会对其进行加密。可搜索加密(SE)允许云服务器根据用户指定的请求以加密的形式管理和搜索数据。然而,强制攻击很少被考虑,在这种情况下,用户可能被迫打开搜索记录和结果。因此,针对强制攻击提出了可否认的SE解决方案,但这些解决方案的存储开销较大,或者没有考虑到对数据所有者和数据用户双方的双重强制情况。本文将遗忘交叉标签协议(OXT)和可否认加密粗略地结合起来,提出了一种可否认SE(可否认交叉标签,DXT)方案,该方案支持布尔查询并抵抗双重强制攻击。从技术上讲,我们形式化了一种称为可更新可否认加密的新原语,并以一种非凡的方式将其与OXT结合起来。此外,给出了DXT的形式化系统模型、安全模型和安全性证明。我们利用华为云平台,基于公共数据集,对DXT和最先进的解决方案进行了充分的对比实验。实验结果表明,DXT在获得更好的特征的同时具有更高的搜索效率。
{"title":"Refrain From Inquiring About My Scalable Storage and Boolean Queries for Secure Cloud","authors":"Boli Hu;Kai Zhang;Junqing Gong;Haifeng Qian","doi":"10.1109/TCC.2025.3582645","DOIUrl":"https://doi.org/10.1109/TCC.2025.3582645","url":null,"abstract":"Outsourcing personal data to a convenient and affordable cloud platform has become a popular practice. Considering the risk of privacy leakage, users usually encrypt their data before uploading it to the cloud server. Searchable encryption (SE) allows cloud servers to manage and search data in encrypted form based on user-specified requests. However, coercion attacks are rarely considered, where users may be forced to open search records and results. Therefore, deniable SE solutions against coercion attacks are presented, but they suffer from large storage overhead or fail to consider the dual coercion situation towards both sides of data owners and data users. In this paper, we roughly combine oblivious cross-tags protocol (OXT) and deniable encryption to propose a deniable SE (deniable cross-tag, DXT) scheme, which supports boolean queries and resists dual coercion attacks. Technically, we formalize a new primitive called updatable deniable encryption, and combine it with OXT in a non-trivial manner. In addition, we give formal system model, security model, and security proof of DXT. By employing the HUAWEI cloud platform, we conduct sufficient comparative experiments between DXT and state-of-the-art solutions based on a public dataset. The experimental results demonstrate that DXT outperforms higher search efficiency while achieving better features.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 3","pages":"969-982"},"PeriodicalIF":5.0,"publicationDate":"2025-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144998092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
REE-TM: Reliable and Energy-Efficient Traffic Management Model for Diverse Cloud Workloads REE-TM:适用于各种云工作负载的可靠、节能的流量管理模型
IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-06-20 DOI: 10.1109/TCC.2025.3581697
Ashutosh Kumar Singh;Deepika Saxena;Volker Lindenstruth
Diversity of workload demands lays a critical impact on efficient resource allocation and management of cloud services. The existing literature has either weakly considered or overlooked the heterogeneous feature of job requests received from wide range of internet services users. To address this context, the proposed approach named Reliable and Energy Efficient Traffic Management (REE-TM) has exploited the diversity of internet traffic in terms of variation in resource demands and expected complexity. Specifically, REE-TM incorporates categorization of heterogeneous job requests and executes them by selecting the most admissible virtual node (a software-defined instance such as a virtual machine or container) and physical node (an actual hardware server or compute host) within the cloud infrastructure. To deal with resource-contention-based resource failures and performance degradation, a novel workload estimator ‘Toffoli Gate-based Quantum Neural Network’ (TG-QNN) is proposed, wherein learning process or interconnection weights optimization is achieved using Quantum version of BlackHole (QBHO) algorithm. The proactively estimated workload is used to compute entropy of the upcoming internet traffic with various traffic states analysis for detection of probable resource-congestion. REE-TM is extensively evaluated through simulations using a benchmark dataset and compared with optimal and without REE-TM versions. The performance evaluation and comparison of REE-TM with measured significant metrics reveal its effectiveness in assuring higher reliability by up to 30.25% and energy-efficiency by up to 23% as compared without REE-TM.
工作负载需求的多样性对云服务的有效资源分配和管理有着至关重要的影响。现有文献要么弱考虑或忽视了从广泛的互联网服务用户收到的工作请求的异构特征。为了解决这一问题,提出了一种名为可靠和节能交通管理(REE-TM)的方法,该方法利用了互联网流量在资源需求和预期复杂性方面的多样性。具体来说,REE-TM结合了异构作业请求的分类,并通过在云基础设施中选择最可接受的虚拟节点(软件定义的实例,如虚拟机或容器)和物理节点(实际的硬件服务器或计算主机)来执行它们。为了解决基于资源竞争的资源故障和性能下降问题,提出了一种基于Toffoli门的量子神经网络(TG-QNN),其中使用量子版黑洞(QBHO)算法实现学习过程或互连权优化。利用主动估计的工作负载计算即将到来的互联网流量的熵,并对各种流量状态进行分析,以检测可能的资源拥塞。通过使用基准数据集的模拟对REE-TM进行了广泛的评估,并与最佳版本和无REE-TM版本进行了比较。REE-TM的性能评估和与实测显著指标的比较表明,与不使用REE-TM相比,其可靠性可提高30.25%,能效可提高23%。
{"title":"REE-TM: Reliable and Energy-Efficient Traffic Management Model for Diverse Cloud Workloads","authors":"Ashutosh Kumar Singh;Deepika Saxena;Volker Lindenstruth","doi":"10.1109/TCC.2025.3581697","DOIUrl":"https://doi.org/10.1109/TCC.2025.3581697","url":null,"abstract":"Diversity of workload demands lays a critical impact on efficient resource allocation and management of cloud services. The existing literature has either weakly considered or overlooked the heterogeneous feature of job requests received from wide range of internet services users. To address this context, the proposed approach named <bold>R</b>eliable and <bold>E</b>nergy <bold>E</b>fficient <bold>T</b>raffic <bold>M</b>anagement (<bold>REE-TM</b>) has exploited the diversity of internet traffic in terms of variation in resource demands and expected complexity. Specifically, REE-TM incorporates categorization of heterogeneous job requests and executes them by selecting the most admissible <italic>virtual node</i> (a software-defined instance such as a virtual machine or container) and <italic>physical node</i> (an actual hardware server or compute host) within the cloud infrastructure. To deal with resource-contention-based resource failures and performance degradation, a novel workload estimator ‘Toffoli Gate-based Quantum Neural Network’ (TG-QNN) is proposed, wherein learning process or interconnection weights optimization is achieved using Quantum version of BlackHole (QBHO) algorithm. The proactively estimated workload is used to compute entropy of the upcoming internet traffic with various traffic states analysis for detection of probable resource-congestion. REE-TM is extensively evaluated through simulations using a benchmark dataset and compared with optimal and without REE-TM versions. The performance evaluation and comparison of REE-TM with measured significant metrics reveal its effectiveness in assuring higher reliability by up to 30.25% and energy-efficiency by up to 23% as compared without REE-TM.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 3","pages":"953-968"},"PeriodicalIF":5.0,"publicationDate":"2025-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144997133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Reference Architecture for Governance of Cloud Native Applications 云原生应用程序治理的参考架构
IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-06-10 DOI: 10.1109/TCC.2025.3578557
William Pourmajidi;Lei Zhang;John Steinbacher;Tony Erwin;Andriy Miranskyy
The evolution of cloud computing has given rise to Cloud Native Applications (CNAs), presenting new challenges in governance, particularly when faced with strict compliance requirements. This work explores the unique characteristics of CNAs and their impact on governance. We introduce a comprehensive reference architecture designed to streamline governance across CNAs, along with a sample implementation, offering insights for both single and multi-cloud environments. Our architecture seamlessly integrates governance within the CNA framework, adhering to a “battery-included” philosophy. Tailored for both expansive and compact CNA deployments across various industries, this design enables cloud practitioners to prioritize product development by alleviating the complexities associated with governance. In addition, it provides a building block for academic exploration of generic CNA frameworks, highlighting their relevance in the evolving cloud computing landscape.
云计算的发展产生了云原生应用程序(CNAs),这给治理带来了新的挑战,特别是在面临严格的遵从性要求时。本研究探讨了中央情报局的独特特征及其对治理的影响。我们介绍了一个全面的参考架构,旨在简化跨cna的治理,以及一个示例实现,为单云和多云环境提供见解。我们的架构在CNA框架内无缝地集成了治理,坚持“电池包括”的理念。针对不同行业的扩展和紧凑的CNA部署,这种设计使云计算从业者能够通过减轻与治理相关的复杂性来确定产品开发的优先级。此外,它还为通用CNA框架的学术探索提供了一个构建块,突出了它们在不断发展的云计算领域中的相关性。
{"title":"A Reference Architecture for Governance of Cloud Native Applications","authors":"William Pourmajidi;Lei Zhang;John Steinbacher;Tony Erwin;Andriy Miranskyy","doi":"10.1109/TCC.2025.3578557","DOIUrl":"https://doi.org/10.1109/TCC.2025.3578557","url":null,"abstract":"The evolution of cloud computing has given rise to Cloud Native Applications (CNAs), presenting new challenges in governance, particularly when faced with strict compliance requirements. This work explores the unique characteristics of CNAs and their impact on governance. We introduce a comprehensive reference architecture designed to streamline governance across CNAs, along with a sample implementation, offering insights for both single and multi-cloud environments. Our architecture seamlessly integrates governance within the CNA framework, adhering to a “battery-included” philosophy. Tailored for both expansive and compact CNA deployments across various industries, this design enables cloud practitioners to prioritize product development by alleviating the complexities associated with governance. In addition, it provides a building block for academic exploration of generic CNA frameworks, highlighting their relevance in the evolving cloud computing landscape.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 3","pages":"935-952"},"PeriodicalIF":5.0,"publicationDate":"2025-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144998014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PHOENIX: Misconfiguration Detection for AWS Serverless Computing PHOENIX:用于AWS无服务器计算的错误配置检测
IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-06-06 DOI: 10.1109/TCC.2025.3577211
Jinfeng Wen;Haodi Ping
Serverless computing is a burgeoning cloud computing paradigm that allows developers to implement applications at the function level, known as serverless applications. Amazon Web Services (AWS), the leading provider in this field, offers Serverless Application Model (AWS SAM), a widely adopted configuration schema for configuring functions and managing resources. However, misconfigurations pose a major challenge during serverless application development, and existing methods are not applicable. To our knowledge, the configuration characteristics and misconfiguration detection for serverless applications have not been well explored. To address this gap, we collect and analyze 733 real-world serverless application configuration files using AWS SAM to understand their characteristics and challenges. Based on the insights, we design PHOENIX, a misconfiguration detection approach for serverless computing. PHOENIX learns configuration patterns from uniform representations of configurations and identifies potential misconfigurations that deviate from these patterns. To evaluate PHOENIX, we construct a dataset comprising 35 injected misconfigurations and 70 real-world misconfigurations with confirmed causes. Our results show that PHOENIX detects 100% of the injected misconfigurations and identifies 97.14% of real-world misconfigurations, significantly outperforming the state-of-the-art tool.
无服务器计算是一种新兴的云计算范式,它允许开发人员在功能级别实现应用程序,称为无服务器应用程序。该领域的领先提供商Amazon Web Services (AWS)提供无服务器应用程序模型(AWS SAM),这是一种广泛采用的配置模式,用于配置功能和管理资源。然而,在无服务器应用程序开发过程中,错误配置是一个主要的挑战,现有的方法不适用。据我们所知,无服务器应用程序的配置特征和错误配置检测还没有得到很好的研究。为了解决这一差距,我们使用AWS SAM收集并分析了733个实际的无服务器应用程序配置文件,以了解它们的特征和挑战。基于这些见解,我们设计了PHOENIX,一种用于无服务器计算的错误配置检测方法。PHOENIX从配置的统一表示中学习配置模式,并识别偏离这些模式的潜在错误配置。为了评估PHOENIX,我们构建了一个包含35个注入错误配置和70个真实错误配置的数据集,这些错误配置具有确定的原因。结果表明,PHOENIX可以检测到100%的注入错误配置,并识别出97.14%的实际错误配置,明显优于最先进的工具。
{"title":"PHOENIX: Misconfiguration Detection for AWS Serverless Computing","authors":"Jinfeng Wen;Haodi Ping","doi":"10.1109/TCC.2025.3577211","DOIUrl":"https://doi.org/10.1109/TCC.2025.3577211","url":null,"abstract":"Serverless computing is a burgeoning cloud computing paradigm that allows developers to implement applications at the function level, known as serverless applications. Amazon Web Services (AWS), the leading provider in this field, offers Serverless Application Model (AWS SAM), a widely adopted configuration schema for configuring functions and managing resources. However, misconfigurations pose a major challenge during serverless application development, and existing methods are not applicable. To our knowledge, the configuration characteristics and misconfiguration detection for serverless applications have not been well explored. To address this gap, we collect and analyze 733 real-world serverless application configuration files using AWS SAM to understand their characteristics and challenges. Based on the insights, we design <italic>PHOENIX</i>, a misconfiguration detection approach for serverless computing. <italic>PHOENIX</i> learns configuration patterns from uniform representations of configurations and identifies potential misconfigurations that deviate from these patterns. To evaluate <italic>PHOENIX</i>, we construct a dataset comprising 35 injected misconfigurations and 70 real-world misconfigurations with confirmed causes. Our results show that <italic>PHOENIX</i> detects 100% of the injected misconfigurations and identifies 97.14% of real-world misconfigurations, significantly outperforming the state-of-the-art tool.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 3","pages":"922-934"},"PeriodicalIF":5.0,"publicationDate":"2025-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144998012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Cross-Workload Power Prediction Method Based on Transfer Gaussian Process Regression in Cloud Data Centers 基于传递高斯过程回归的云数据中心跨工作负载功率预测方法
IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-06-02 DOI: 10.1109/TCC.2025.3575790
Ruichao Mo;Weiwei Lin;Haocheng Zhong;Minxian Xu;Keqin Li
Nowadays, machine learning (ML)-based power prediction models for servers have shown remarkable performance, leveraging large volumes of labeled data for training. However, collecting extensive labeled power data from servers in cloud data centers incurs substantial costs. Additionally, varying resource demands across different workloads (e.g., CPU-intensive, memory-intensive, and I/O-intensive) lead to significant differences in power consumption behaviors, known as domain shift. Consequently, power data collected from one type of workload cannot effectively train power prediction models for other workloads, limiting the exploration of the collected power data. To tackle these challenges, we propose TGCP, a cross-workload power prediction method based on multi-source transfer Gaussian process regression. TGCP transfers knowledge from abundant power data across multiple source workloads to a target workload with limited power data. Furthermore, Continuous normalizing flows adjust the posterior prediction distribution of Gaussian process, making it locally non-Gaussian, enhancing TGCP’s ability to handle real-world power data distribution. This method enhances prediction accuracy for the target workload while reducing the expense of acquiring power data for real cloud data centers. Experimental results on a realistic power consumption dataset demonstrate that TGCP surpasses four traditional ML methods and three transfer learning methods in cross-workload power prediction.
如今,基于机器学习(ML)的服务器功率预测模型已经显示出卓越的性能,利用大量标记数据进行训练。然而,从云数据中心的服务器收集大量标记电源数据会产生大量成本。此外,跨不同工作负载(例如,cpu密集型、内存密集型和I/ o密集型)的不同资源需求会导致功耗行为的显著差异,称为域转移。因此,从一种工作负载中收集的功率数据不能有效地训练用于其他工作负载的功率预测模型,从而限制了对所收集的功率数据的探索。为了解决这些问题,我们提出了一种基于多源传递高斯过程回归的跨工作负载功率预测方法TGCP。TGCP将知识从跨多个源工作负载的丰富电力数据传输到具有有限电力数据的目标工作负载。此外,连续归一化流调整高斯过程的后验预测分布,使其局部非高斯分布,增强了TGCP处理实际功率数据分布的能力。该方法提高了对目标工作负载的预测精度,同时减少了获取真实云数据中心电力数据的费用。在实际功耗数据集上的实验结果表明,TGCP在跨工作负载功耗预测方面优于4种传统ML方法和3种迁移学习方法。
{"title":"A Cross-Workload Power Prediction Method Based on Transfer Gaussian Process Regression in Cloud Data Centers","authors":"Ruichao Mo;Weiwei Lin;Haocheng Zhong;Minxian Xu;Keqin Li","doi":"10.1109/TCC.2025.3575790","DOIUrl":"https://doi.org/10.1109/TCC.2025.3575790","url":null,"abstract":"Nowadays, machine learning (ML)-based power prediction models for servers have shown remarkable performance, leveraging large volumes of labeled data for training. However, collecting extensive labeled power data from servers in cloud data centers incurs substantial costs. Additionally, varying resource demands across different workloads (e.g., CPU-intensive, memory-intensive, and I/O-intensive) lead to significant differences in power consumption behaviors, known as domain shift. Consequently, power data collected from one type of workload cannot effectively train power prediction models for other workloads, limiting the exploration of the collected power data. To tackle these challenges, we propose <italic>TGCP</i>, a cross-workload power prediction method based on multi-source transfer Gaussian process regression. <italic>TGCP</i> transfers knowledge from abundant power data across multiple source workloads to a target workload with limited power data. Furthermore, Continuous normalizing flows adjust the posterior prediction distribution of Gaussian process, making it locally non-Gaussian, enhancing <italic>TGCP</i>’s ability to handle real-world power data distribution. This method enhances prediction accuracy for the target workload while reducing the expense of acquiring power data for real cloud data centers. Experimental results on a realistic power consumption dataset demonstrate that <italic>TGCP</i> surpasses four traditional ML methods and three transfer learning methods in cross-workload power prediction.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 3","pages":"910-921"},"PeriodicalIF":5.0,"publicationDate":"2025-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144997970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing Cloud Computing Performance Through Integration of a Threshold-Based Load Balancing Algorithm With Multiple Service Broker Policies 通过集成基于阈值的负载均衡算法和多个服务代理策略来优化云计算性能
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-04-24 DOI: 10.1109/TCC.2025.3563848
Shusmoy Chowdhury;Ajay Katangur
The triumph of cloud computing hinges upon the adept instantiation of infrastructure and the judicious utilization of available resources. Load balancing, a pivotal facet, substantiates the fulfillment of these imperatives, thereby augmenting the performance of the cloud environment for its users. Our research introduces a load balancing algorithm grounded in threshold principles devised to ensure equitable distribution of workloads among nodes. The main objective of the algorithm is to preclude the overburdening of virtual machines (VMs) within the cloud with tasks or their idleness due to task allocation deficiencies in the presence of active tasks. The threshold values embedded in our algorithm ascertain the judicious deployment of VMs, forestalling both task overload and idle states arising from task allocation inadequacies. Simulation outcomes manifest that our threshold-based algorithm markedly enhances response time for tasks/requests and data processing duration within datacenters, outperforming extant algorithms such as First Come First Serve, Round Robin, and the Equally Spread Current Execution Load Balancing algorithm. Our threshold algorithm attains superior results to alternative load balancing algorithms when coupled with an optimized response time service broker policy.
云计算的成功取决于基础设施的熟练实例化和对可用资源的明智利用。负载平衡是一个关键方面,它证实了这些需求的实现,从而为用户增强了云环境的性能。我们的研究引入了一种基于阈值原则的负载平衡算法,旨在确保节点之间公平分配工作负载。该算法的主要目标是防止云中的虚拟机(vm)因任务过重或在活动任务存在时由于任务分配不足而导致的空闲。我们的算法中嵌入的阈值确定了vm的明智部署,防止了任务过载和由于任务分配不足而引起的空闲状态。模拟结果表明,我们的基于阈值的算法显著提高了数据中心内任务/请求的响应时间和数据处理持续时间,优于现有的算法,如先到先服务、轮询和平均分布当前执行负载平衡算法。当与优化的响应时间服务代理策略相结合时,我们的阈值算法比其他负载平衡算法获得更好的结果。
{"title":"Optimizing Cloud Computing Performance Through Integration of a Threshold-Based Load Balancing Algorithm With Multiple Service Broker Policies","authors":"Shusmoy Chowdhury;Ajay Katangur","doi":"10.1109/TCC.2025.3563848","DOIUrl":"https://doi.org/10.1109/TCC.2025.3563848","url":null,"abstract":"The triumph of cloud computing hinges upon the adept instantiation of infrastructure and the judicious utilization of available resources. Load balancing, a pivotal facet, substantiates the fulfillment of these imperatives, thereby augmenting the performance of the cloud environment for its users. Our research introduces a load balancing algorithm grounded in threshold principles devised to ensure equitable distribution of workloads among nodes. The main objective of the algorithm is to preclude the overburdening of virtual machines (VMs) within the cloud with tasks or their idleness due to task allocation deficiencies in the presence of active tasks. The threshold values embedded in our algorithm ascertain the judicious deployment of VMs, forestalling both task overload and idle states arising from task allocation inadequacies. Simulation outcomes manifest that our threshold-based algorithm markedly enhances response time for tasks/requests and data processing duration within datacenters, outperforming extant algorithms such as First Come First Serve, Round Robin, and the Equally Spread Current Execution Load Balancing algorithm. Our threshold algorithm attains superior results to alternative load balancing algorithms when coupled with an optimized response time service broker policy.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 2","pages":"751-768"},"PeriodicalIF":5.3,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144232134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modeling Resource Scheduling in Optical Switching DCNs Under Bursty and Skewed Traffic 突发和倾斜流量下光交换DCNs资源调度建模
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-04-16 DOI: 10.1109/TCC.2025.3561281
Shuai Zhang;Baojun Chen;Weiqiang Sun;Weisheng Hu
When optical switching is deployed in Data Center Networks (DCNs), the reconfiguration of the optical switching matrix leads to substantially longer overheads, posing a significant impact on the system performance. Despite the extensive studies on the scheduling algorithms based on demand matrix decomposition (DMD), the stateful and irregular nature of the scheduling processes hinders the development of quantitative models, thereby limiting our understanding of resource scheduling in optical switching DCNs based on DMD. In this article, we model the DMD based resource scheduling process under a bursty and skewed traffic pattern and derive closed-form equations for the burst completion time. Our study shows that an increased reconfiguration delay will lead to an approximate linear increase in the burst completion time. Our study also demonstrates that the size of the slot and the maximum allowed duration of one match are approximately inversely proportional to the burst completion time, with diminishing marginal returns.
当光交换部署在数据中心网络(DCNs)中时,光交换矩阵的重新配置将导致开销大幅增加,对系统性能产生重大影响。尽管基于需求矩阵分解(DMD)的调度算法得到了广泛的研究,但调度过程的状态性和不规则性阻碍了定量模型的发展,从而限制了我们对基于DMD的光交换dns资源调度的理解。在本文中,我们建立了基于DMD的资源调度过程在突发和倾斜交通模式下的模型,并推导了突发完成时间的封闭形式方程。我们的研究表明,重构延迟的增加将导致爆破完井时间的近似线性增加。我们的研究还表明,狭缝的大小和一次匹配的最大允许持续时间与突发完成时间近似成反比,边际收益递减。
{"title":"Modeling Resource Scheduling in Optical Switching DCNs Under Bursty and Skewed Traffic","authors":"Shuai Zhang;Baojun Chen;Weiqiang Sun;Weisheng Hu","doi":"10.1109/TCC.2025.3561281","DOIUrl":"https://doi.org/10.1109/TCC.2025.3561281","url":null,"abstract":"When optical switching is deployed in Data Center Networks (DCNs), the reconfiguration of the optical switching matrix leads to substantially longer overheads, posing a significant impact on the system performance. Despite the extensive studies on the scheduling algorithms based on demand matrix decomposition (DMD), the stateful and irregular nature of the scheduling processes hinders the development of quantitative models, thereby limiting our understanding of resource scheduling in optical switching DCNs based on DMD. In this article, we model the DMD based resource scheduling process under a bursty and skewed traffic pattern and derive closed-form equations for the burst completion time. Our study shows that an increased reconfiguration delay will lead to an approximate linear increase in the burst completion time. Our study also demonstrates that the size of the slot and the maximum allowed duration of one match are approximately inversely proportional to the burst completion time, with diminishing marginal returns.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 2","pages":"737-750"},"PeriodicalIF":5.3,"publicationDate":"2025-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144230587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Secure kNN for Distributed Cloud Environment Using Fully Homomorphic Encryption 基于全同态加密的分布式云环境安全kNN
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-04-16 DOI: 10.1109/TCC.2025.3561586
Yuuya Fukuchi;Sota Hashimoto;Kazuya Sakai;Satoshi Fukumoto;Min-Te Sun;Wei-Shinn Ku
Privacy-preserving k-nearest neighbor (PPkNN) classification for multiple clouds enables categorizing queried data into a class in keeping with data privacy, where the database and key servers jointly perform cryptographic operations. The existing solutions, unfortunately, take a long time and incur a large amount of traffic between the database and key servers. Therefore, in this article, we propose a fast and secure kNN classification protocol, namely FSkNN, over distributed databases deployed in multiple clouds under the semi-honest model. Particularly, we focus on optimizing the network-related operations during kNN classification. That is, the proposed cryptographic protocol reduces the number of interactions between the servers by using a fully homomorphic encryption scheme and eliminates unnecessary traffic by applying mathematical techniques. In addition, the indistinguishability-based security of FSkNN is proven. We implemented FSkNN with C++ and the testbed experiments demonstrate that the proposed scheme significantly facilitates the query response time and reduces the communication cost.
针对多个云的保护隐私的k-最近邻(PPkNN)分类支持将查询的数据分类到符合数据隐私的类中,其中数据库和密钥服务器联合执行加密操作。不幸的是,现有的解决方案耗时很长,并且在数据库和密钥服务器之间产生大量流量。因此,在本文中,我们在半诚实模型下,针对部署在多云中的分布式数据库,提出了一种快速安全的kNN分类协议,即FSkNN。特别是,我们专注于优化kNN分类过程中与网络相关的操作。也就是说,所提出的加密协议通过使用完全同态的加密方案减少了服务器之间的交互次数,并通过应用数学技术消除了不必要的流量。此外,验证了FSkNN基于不可区分性的安全性。用c++实现了FSkNN,实验结果表明,该方案显著缩短了查询响应时间,降低了通信成本。
{"title":"Secure kNN for Distributed Cloud Environment Using Fully Homomorphic Encryption","authors":"Yuuya Fukuchi;Sota Hashimoto;Kazuya Sakai;Satoshi Fukumoto;Min-Te Sun;Wei-Shinn Ku","doi":"10.1109/TCC.2025.3561586","DOIUrl":"https://doi.org/10.1109/TCC.2025.3561586","url":null,"abstract":"Privacy-preserving k-nearest neighbor (PPkNN) classification for multiple clouds enables categorizing queried data into a class in keeping with data privacy, where the database and key servers jointly perform cryptographic operations. The existing solutions, unfortunately, take a long time and incur a large amount of traffic between the database and key servers. Therefore, in this article, we propose a fast and secure kNN classification protocol, namely FSkNN, over distributed databases deployed in multiple clouds under the semi-honest model. Particularly, we focus on optimizing the network-related operations during kNN classification. That is, the proposed cryptographic protocol reduces the number of interactions between the servers by using a fully homomorphic encryption scheme and eliminates unnecessary traffic by applying mathematical techniques. In addition, the indistinguishability-based security of FSkNN is proven. We implemented FSkNN with C++ and the testbed experiments demonstrate that the proposed scheme significantly facilitates the query response time and reduces the communication cost.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"13 2","pages":"721-736"},"PeriodicalIF":5.3,"publicationDate":"2025-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144230588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Cloud Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1