首页 > 最新文献

IEEE Transactions on Cloud Computing最新文献

英文 中文
QoS-Aware, Cost-Efficient Scheduling for Data-Intensive DAGs in Multi-Tier Computing Environment 多层计算环境下数据密集型dag的qos感知、成本高效调度
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-26 DOI: 10.1109/TCC.2024.3468913
Paridhika Kayal;Alberto Leon-Garcia
In today’s scientific landscape, Directed Acyclic Graphs (DAGs) are pivotal for representing task dependencies in data-intensive applications. Traditionally, two dominant bottom-up DAG scheduling approaches exist: one overlooks communication contention and the other fails to exploit parallelization for improving latency. This study distinguishes itself by advocating a top-down approach prioritizing latency or cost optimization in multi-tier environments to fulfill QoS and SLA requirements. Our strategy effectively mitigates bandwidth contention and facilitates parallel executions, leading to substantial completion time reductions. Our findings suggest that myopic knowledge-based scheduling, emphasizing latency or cost minimization, can yield benefits comparable to its look-ahead counterparts. Through latency-efficient and cost-efficient topological sorting, our wDAGSplit strategy introduces a two-stage partitioning and scheduling approach. Its simplicity and adaptability extend its usability to DAGs of any scale. Evaluated on over 100,000 real-world DAG applications, wDAGSplit demonstrates latency enhancements of up to 80x compared to Edge-only scenarios, 15x to Near-Edge-only, and 6x to Cloud-only. In terms of cost, our approach achieves enhancements of up to 60x compared to Edge-only scenarios, 250x to NE-only, and 70x to Cloud-only. Moreover, for DAGs with 50 tasks, we achieve 5x reduced latency compared to previous approaches, along with a complexity reduction of up to 24 times.
在当今的科学领域,有向无环图(dag)对于表示数据密集型应用程序中的任务依赖关系至关重要。传统上,存在两种主要的自底向上DAG调度方法:一种忽略通信争用,另一种没有利用并行化来改善延迟。本研究的独特之处在于提倡在多层环境中优先考虑延迟或成本优化的自上而下方法,以满足QoS和SLA要求。我们的策略有效地缓解了带宽争用,促进了并行执行,从而大大减少了完成时间。我们的研究结果表明,短视的基于知识的调度,强调延迟或成本最小化,可以产生与前瞻性调度相当的好处。通过延迟高效和成本高效的拓扑排序,我们的wDAGSplit策略引入了一种两阶段分区和调度方法。它的简单性和适应性将其可用性扩展到任何规模的dag。在超过100,000个真实的DAG应用程序上进行评估后,wDAGSplit显示,与仅边缘场景相比,延迟增强高达80倍,与仅近边缘场景相比增强15倍,与仅云场景相比增强6倍。在成本方面,我们的方法与Edge-only方案相比可实现高达60倍的增强,与NE-only方案相比可实现250倍的增强,与Cloud-only方案相比可实现70倍的增强。此外,对于具有50个任务的dag,与以前的方法相比,我们实现了5倍的延迟降低,以及高达24倍的复杂性降低。
{"title":"QoS-Aware, Cost-Efficient Scheduling for Data-Intensive DAGs in Multi-Tier Computing Environment","authors":"Paridhika Kayal;Alberto Leon-Garcia","doi":"10.1109/TCC.2024.3468913","DOIUrl":"https://doi.org/10.1109/TCC.2024.3468913","url":null,"abstract":"In today’s scientific landscape, Directed Acyclic Graphs (DAGs) are pivotal for representing task dependencies in data-intensive applications. Traditionally, two dominant bottom-up DAG scheduling approaches exist: one overlooks communication contention and the other fails to exploit parallelization for improving latency. This study distinguishes itself by advocating a top-down approach prioritizing latency or cost optimization in multi-tier environments to fulfill QoS and SLA requirements. Our strategy effectively mitigates bandwidth contention and facilitates parallel executions, leading to substantial completion time reductions. Our findings suggest that myopic knowledge-based scheduling, emphasizing latency or cost minimization, can yield benefits comparable to its look-ahead counterparts. Through latency-efficient and cost-efficient topological sorting, our \u0000<italic>wDAGSplit</i>\u0000 strategy introduces a two-stage partitioning and scheduling approach. Its simplicity and adaptability extend its usability to DAGs of any scale. Evaluated on over 100,000 real-world DAG applications, \u0000<italic>wDAGSplit</i>\u0000 demonstrates latency enhancements of up to 80x compared to Edge-only scenarios, 15x to Near-Edge-only, and 6x to Cloud-only. In terms of cost, our approach achieves enhancements of up to 60x compared to Edge-only scenarios, 250x to NE-only, and 70x to Cloud-only. Moreover, for DAGs with 50 tasks, we achieve 5x reduced latency compared to previous approaches, along with a complexity reduction of up to 24 times.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 4","pages":"1314-1327"},"PeriodicalIF":5.3,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142777763","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anomaly Transformer Ensemble Model for Cloud Data Anomaly Detection 云数据异常检测的异常变压器集成模型
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-23 DOI: 10.1109/TCC.2024.3466174
Won Sakong;Jongyeop Kwon;Kyungha Min;Suyeon Wang;Wooju Kim
The stability and user trust in cloud services depends on prompt detection and response to diverse anomalies. This study focuses on an Ensemble-based anomaly detection methodology that integrates log data with computing resource metrics, aiming to overcome the limitations of traditional single-data models. To process the unstructured nature of log data, we use the Drain Parser to transform it into a structured format, and Doc2Vec embeds it. The study adheres to a reconstruction-based approach for anomaly detection, specifically building upon the Anomaly Transformer model. The proposed model leverages the concept of an Anomaly Transformer based on the Attention mechanism. It integrates preprocessed metric data with log data for effective anomaly detection. Experiments were conducted using metric and log data collected from real-world cloud environments. The model’s performance was evaluated based on accuracy, recall, precision, f1 score, and AUROC. The results demonstrate that our proposed Ensemble-based model outperforms traditional models such as LSTM, VAR, and deeplog.
云服务的稳定性和用户信任度取决于对各种异常的及时检测和响应。本研究的重点是基于集成的异常检测方法,该方法将测井数据与计算资源指标相结合,旨在克服传统单一数据模型的局限性。为了处理日志数据的非结构化特性,我们使用Drain Parser将其转换为结构化格式,然后Doc2Vec将其嵌入。该研究坚持基于重构的异常检测方法,特别是建立在异常变压器模型之上。提出的模型利用了基于注意力机制的异常转换器的概念。它将预处理的度量数据与日志数据相结合,有效地进行异常检测。实验使用从真实的云环境中收集的度量和日志数据进行。模型的性能根据准确率、召回率、精度、f1分数和AUROC进行评估。结果表明,我们提出的基于集成的模型优于LSTM、VAR和deeplog等传统模型。
{"title":"Anomaly Transformer Ensemble Model for Cloud Data Anomaly Detection","authors":"Won Sakong;Jongyeop Kwon;Kyungha Min;Suyeon Wang;Wooju Kim","doi":"10.1109/TCC.2024.3466174","DOIUrl":"https://doi.org/10.1109/TCC.2024.3466174","url":null,"abstract":"The stability and user trust in cloud services depends on prompt detection and response to diverse anomalies. This study focuses on an Ensemble-based anomaly detection methodology that integrates log data with computing resource metrics, aiming to overcome the limitations of traditional single-data models. To process the unstructured nature of log data, we use the Drain Parser to transform it into a structured format, and Doc2Vec embeds it. The study adheres to a reconstruction-based approach for anomaly detection, specifically building upon the Anomaly Transformer model. The proposed model leverages the concept of an Anomaly Transformer based on the Attention mechanism. It integrates preprocessed metric data with log data for effective anomaly detection. Experiments were conducted using metric and log data collected from real-world cloud environments. The model’s performance was evaluated based on accuracy, recall, precision, f1 score, and AUROC. The results demonstrate that our proposed Ensemble-based model outperforms traditional models such as LSTM, VAR, and deeplog.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 4","pages":"1305-1313"},"PeriodicalIF":5.3,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142777685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
WorkloadDiff: Conditional Denoising Diffusion Probabilistic Models for Cloud Workload Prediction WorkloadDiff:用于云计算工作量预测的条件去噪扩散概率模型
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-16 DOI: 10.1109/TCC.2024.3461649
Weiping Zheng;Zongxiao Chen;Kaiyuan Zheng;Weijian Zheng;Yiqi Chen;Xiaomao Fan
Accurate workload forecasting plays a crucial role in optimizing resource allocation, enhancing performance, and reducing energy consumption in cloud data centers. Deep learning-based methods have emerged as the dominant approach in this field, exhibiting exceptional performance. However, most existing methods lack the ability to quantify confidence, limiting their practical decision-making utility. To address this limitation, we propose a novel denoising diffusion probabilistic model (DDPM)-based method, termed WorkloadDiff, for multivariate probabilistic workload prediction. WorkloadDiff leverages both original and noisy signals from input conditions using a two-path neural network. Additionally, we introduce a multi-scale feature extraction method and an adaptive fusion approach to capture diverse temporal patterns within the workload. To enhance consistency between conditions and predicted values, we incorporate a resampling strategy into the inference of WorkloadDiff. Extensive experiments conducted on four public datasets demonstrate the superior performance of WorkloadDiff over all baseline models, establishing it as a robust tool for resource management in cloud data centers.
准确的工作负载预测在云数据中心优化资源分配、提高性能、降低能耗方面发挥着至关重要的作用。基于深度学习的方法已经成为该领域的主导方法,表现出优异的性能。然而,大多数现有方法缺乏量化信心的能力,限制了它们的实际决策效用。为了解决这一限制,我们提出了一种新的基于扩散概率模型(DDPM)的去噪方法,称为WorkloadDiff,用于多变量概率工作负载预测。workloadff利用双路径神经网络从输入条件中获取原始信号和噪声信号。此外,我们还引入了一种多尺度特征提取方法和一种自适应融合方法来捕获工作负载中的不同时间模式。为了提高条件和预测值之间的一致性,我们在workloadff的推理中加入了重采样策略。在四个公共数据集上进行的大量实验表明,workloadadff优于所有基线模型,使其成为云数据中心资源管理的强大工具。
{"title":"WorkloadDiff: Conditional Denoising Diffusion Probabilistic Models for Cloud Workload Prediction","authors":"Weiping Zheng;Zongxiao Chen;Kaiyuan Zheng;Weijian Zheng;Yiqi Chen;Xiaomao Fan","doi":"10.1109/TCC.2024.3461649","DOIUrl":"10.1109/TCC.2024.3461649","url":null,"abstract":"Accurate workload forecasting plays a crucial role in optimizing resource allocation, enhancing performance, and reducing energy consumption in cloud data centers. Deep learning-based methods have emerged as the dominant approach in this field, exhibiting exceptional performance. However, most existing methods lack the ability to quantify confidence, limiting their practical decision-making utility. To address this limitation, we propose a novel denoising diffusion probabilistic model (DDPM)-based method, termed WorkloadDiff, for multivariate probabilistic workload prediction. WorkloadDiff leverages both original and noisy signals from input conditions using a two-path neural network. Additionally, we introduce a multi-scale feature extraction method and an adaptive fusion approach to capture diverse temporal patterns within the workload. To enhance consistency between conditions and predicted values, we incorporate a resampling strategy into the inference of WorkloadDiff. Extensive experiments conducted on four public datasets demonstrate the superior performance of WorkloadDiff over all baseline models, establishing it as a robust tool for resource management in cloud data centers.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 4","pages":"1291-1304"},"PeriodicalIF":5.3,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142260182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Lightweight Privacy-Preserving Ciphertext Retrieval Scheme Based on Edge Computing 基于边缘计算的轻量级隐私保护密文检索方案
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-16 DOI: 10.1109/TCC.2024.3461732
Na Wang;Wen Zhou;Qingyun Han;Jianwei Liu;Weilue Liao;Junsong Fu
With the rapid development of cloud computing and Internet of Things (IoT) technologies, large amounts of data collected from IoT devices are encrypted and outsourced to cloud servers for storage and sharing. However, traditional ciphertext retrieval schemes impose high computation and storage overhead on end users. Meanwhile, IoT devices with limited resources are difficult to adapt to large amounts of data computation and transmission, which leads to transmission delay and poor user experience. In this article, we propose a lightweight privacy-preserving ciphertext retrieval scheme based on edge computing (LPCR) by extending searchable encryption (SE) and ciphertext policy attribute-based encryption (CP-ABE) techniques. First, to avoid network delay and paralysis, we introduce edge servers into LPCR and design a collaboration mechanism between the user side and the edge servers. The user side only needs to accomplish lightweight computation and storage tasks, which greatly reduces their resource consumption. Second, we extend the basic ciphertext policy attribute-based keyword search (CP-ABKS) technique and design the Linear Secret Sharing Scheme (LSSS) access control algorithm with attribute values to hide access policies and attributes. In addition, to improve the retrieval accuracy, the document indexes and query trapdoors are set up by conjunctive keywords to help the cloud server locate exactly the data that the user wishes to query. Formal security analysis verifies that LPCR can achieve the security of chosen plaintext attack (CPA) and chosen keyword attack (CKA), and resist collusion attack. Simulation experiments prove that LPCR is lightweight and feasible.
随着云计算和物联网技术的快速发展,从物联网设备中收集的大量数据被加密并外包给云服务器进行存储和共享。然而,传统的密文检索方案给终端用户带来了较高的计算和存储开销。同时,资源有限的物联网设备难以适应大量数据的计算和传输,导致传输延迟,用户体验不佳。在本文中,我们通过扩展可搜索加密(SE)和基于密文策略属性的加密(CP-ABE)技术,提出了一种基于边缘计算(LPCR)的轻量级保密密文检索方案。首先,为了避免网络延迟和瘫痪,我们将边缘服务器引入到LPCR中,并设计了用户端与边缘服务器之间的协作机制。用户只需要完成轻量级的计算和存储任务,这大大降低了用户的资源消耗。其次,我们扩展了基本的基于密文策略属性的关键字搜索(CP-ABKS)技术,设计了带有属性值的线性秘密共享方案(LSSS)访问控制算法来隐藏访问策略和属性。此外,为了提高检索精度,通过连接关键词建立文档索引和查询陷阱门,帮助云服务器准确定位用户想要查询的数据。形式化的安全性分析验证了LPCR可以实现选择明文攻击(CPA)和选择关键字攻击(CKA)的安全性,并能抵抗合众攻击。仿真实验证明了LPCR的轻量化和可行性。
{"title":"A Lightweight Privacy-Preserving Ciphertext Retrieval Scheme Based on Edge Computing","authors":"Na Wang;Wen Zhou;Qingyun Han;Jianwei Liu;Weilue Liao;Junsong Fu","doi":"10.1109/TCC.2024.3461732","DOIUrl":"10.1109/TCC.2024.3461732","url":null,"abstract":"With the rapid development of cloud computing and Internet of Things (IoT) technologies, large amounts of data collected from IoT devices are encrypted and outsourced to cloud servers for storage and sharing. However, traditional ciphertext retrieval schemes impose high computation and storage overhead on end users. Meanwhile, IoT devices with limited resources are difficult to adapt to large amounts of data computation and transmission, which leads to transmission delay and poor user experience. In this article, we propose a lightweight privacy-preserving ciphertext retrieval scheme based on edge computing (LPCR) by extending searchable encryption (SE) and ciphertext policy attribute-based encryption (CP-ABE) techniques. First, to avoid network delay and paralysis, we introduce edge servers into LPCR and design a collaboration mechanism between the user side and the edge servers. The user side only needs to accomplish lightweight computation and storage tasks, which greatly reduces their resource consumption. Second, we extend the basic ciphertext policy attribute-based keyword search (CP-ABKS) technique and design the Linear Secret Sharing Scheme (LSSS) access control algorithm with attribute values to hide access policies and attributes. In addition, to improve the retrieval accuracy, the document indexes and query trapdoors are set up by conjunctive keywords to help the cloud server locate exactly the data that the user wishes to query. Formal security analysis verifies that LPCR can achieve the security of chosen plaintext attack (CPA) and chosen keyword attack (CKA), and resist collusion attack. Simulation experiments prove that LPCR is lightweight and feasible.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 4","pages":"1273-1290"},"PeriodicalIF":5.3,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142260183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generative Adversarial Privacy for Multimedia Analytics Across the IoT-Edge Continuum 用于跨物联网-边缘连续性多媒体分析的生成对抗式隐私保护
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-12 DOI: 10.1109/TCC.2024.3459789
Xin Wang;Jianhui Lv;Byung-Gyu Kim;Carsten Maple;B. D. Parameshachari;Adam Slowik;Keqin Li
The proliferation of multimedia-enabled IoT devices and edge computing enables a new class of data-intensive applications. However, analyzing the massive volumes of multimedia data presents significant privacy challenges. We propose a novel framework called generative adversarial privacy (GAP) that leverages generative adversarial networks (GANs) to synthesize privacy-preserving surrogate data for multimedia analytics across the IoT-Edge continuum. GAP carefully perturbs the GAN's training process to provide rigorous differential privacy guarantees without compromising utility. Moreover, we present optimization strategies, including dynamic privacy budget allocation, adaptive gradient clipping, and weight clustering to improve convergence and data quality under a constrained privacy budget. Theoretical analysis proves that GAP provides rigorous privacy protections while enabling high-fidelity analytics. Extensive experiments on real-world multimedia datasets demonstrate that GAP outperforms existing methods, producing high-quality synthetic data for privacy-preserving multimedia processing in diverse IoT-Edge applications.
支持多媒体的物联网设备和边缘计算的激增使一类新的数据密集型应用成为可能。然而,分析大量的多媒体数据带来了重大的隐私挑战。我们提出了一种称为生成对抗隐私(GAP)的新框架,该框架利用生成对抗网络(gan)合成保护隐私的代理数据,用于跨物联网边缘连续体的多媒体分析。GAP仔细地干扰GAN的训练过程,在不影响效用的情况下提供严格的差分隐私保证。此外,我们还提出了动态隐私预算分配、自适应梯度裁剪和权重聚类等优化策略,以提高隐私预算约束下的收敛性和数据质量。理论分析证明,GAP在实现高保真分析的同时提供了严格的隐私保护。在真实世界的多媒体数据集上进行的大量实验表明,GAP优于现有方法,可以为各种物联网边缘应用中的保护隐私的多媒体处理产生高质量的合成数据。
{"title":"Generative Adversarial Privacy for Multimedia Analytics Across the IoT-Edge Continuum","authors":"Xin Wang;Jianhui Lv;Byung-Gyu Kim;Carsten Maple;B. D. Parameshachari;Adam Slowik;Keqin Li","doi":"10.1109/TCC.2024.3459789","DOIUrl":"10.1109/TCC.2024.3459789","url":null,"abstract":"The proliferation of multimedia-enabled IoT devices and edge computing enables a new class of data-intensive applications. However, analyzing the massive volumes of multimedia data presents significant privacy challenges. We propose a novel framework called generative adversarial privacy (GAP) that leverages generative adversarial networks (GANs) to synthesize privacy-preserving surrogate data for multimedia analytics across the IoT-Edge continuum. GAP carefully perturbs the GAN's training process to provide rigorous differential privacy guarantees without compromising utility. Moreover, we present optimization strategies, including dynamic privacy budget allocation, adaptive gradient clipping, and weight clustering to improve convergence and data quality under a constrained privacy budget. Theoretical analysis proves that GAP provides rigorous privacy protections while enabling high-fidelity analytics. Extensive experiments on real-world multimedia datasets demonstrate that GAP outperforms existing methods, producing high-quality synthetic data for privacy-preserving multimedia processing in diverse IoT-Edge applications.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 4","pages":"1260-1272"},"PeriodicalIF":5.3,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142218868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Corrections to “DNN Surgery: Accelerating DNN Inference on the Edge through Layer Partitioning” DNN Surgery:通过层划分加速边缘 DNN 推断" Correct to "DNN Surgery: Accelerating DNN Inference on the Edge through Layer Partitioning"
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-05 DOI: 10.1109/TCC.2024.3404548
Huanghuang Liang;Qianlong Sang;Chuang Hu;Dazhao Cheng;Xiaobo Zhou;Dan Wang;Wei Bao;Yu Wang
In this paper, we reference the previous conference version and complete the grant number mentioned in the acknowledgments of the conference version.
在本文中,我们参考了之前的会议版本,并填写了会议版本致谢中提到的资助编号。
{"title":"Corrections to “DNN Surgery: Accelerating DNN Inference on the Edge through Layer Partitioning”","authors":"Huanghuang Liang;Qianlong Sang;Chuang Hu;Dazhao Cheng;Xiaobo Zhou;Dan Wang;Wei Bao;Yu Wang","doi":"10.1109/TCC.2024.3404548","DOIUrl":"https://doi.org/10.1109/TCC.2024.3404548","url":null,"abstract":"In this paper, we reference the previous conference version and complete the grant number mentioned in the acknowledgments of the conference version.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 3","pages":"966-966"},"PeriodicalIF":5.3,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10666943","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142143671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FedPAW: Federated Learning With Personalized Aggregation Weights for Urban Vehicle Speed Prediction FedPAW:利用个性化聚合权重进行联合学习,用于城市车辆速度预测
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-02 DOI: 10.1109/TCC.2024.3452696
Yuepeng He;Pengzhan Zhou;Yijun Zhai;Fang Qu;Zhida Qin;Mingyan Li;Songtao Guo
Vehicle speed prediction is crucial for intelligent transportation systems, promoting more reliable autonomous driving by accurately predicting future vehicle conditions. Due to variations in drivers’ driving styles and vehicle types, speed predictions for different target vehicles may significantly differ. Existing methods may not realize personalized vehicle speed prediction while protecting drivers’ data privacy. We propose a Federated learning framework with Personalized Aggregation Weights (FedPAW) to overcome these challenges. This method captures client-specific information by measuring the weighted mean squared error between the parameters of local models and global models. The server sends tailored aggregated models to clients instead of a single global model, without incurring additional computational and communication overhead for clients. To evaluate the effectiveness of FedPAW, we collected driving data in urban scenarios using the autonomous driving simulator CARLA, employing an LSTM-based Seq2Seq model with a multi-head attention mechanism to predict the future speed of target vehicles. The results demonstrate that our proposed FedPAW ranks lowest in prediction error within the time horizon of 10 seconds, with a 0.8% reduction in test MAE, compared to eleven representative benchmark baselines.
车速预测对于智能交通系统至关重要,通过准确预测未来车辆状况来促进更可靠的自动驾驶。由于驾驶员驾驶风格和车辆类型的差异,对不同目标车辆的速度预测可能存在显著差异。现有方法在保护驾驶员数据隐私的同时,可能无法实现个性化的车速预测。我们提出了一种具有个性化聚合权重(FedPAW)的联邦学习框架来克服这些挑战。该方法通过测量局部模型和全局模型参数之间的加权均方误差来捕获特定于客户端的信息。服务器向客户端发送定制的聚合模型,而不是单一的全局模型,不会为客户端带来额外的计算和通信开销。为了评估FedPAW的有效性,我们使用自动驾驶模拟器CARLA收集了城市场景下的驾驶数据,采用基于lstm的Seq2Seq模型和多头注意机制来预测目标车辆的未来速度。结果表明,与11个代表性基准基线相比,我们提出的FedPAW在10秒的时间范围内的预测误差最低,测试MAE降低了0.8%。
{"title":"FedPAW: Federated Learning With Personalized Aggregation Weights for Urban Vehicle Speed Prediction","authors":"Yuepeng He;Pengzhan Zhou;Yijun Zhai;Fang Qu;Zhida Qin;Mingyan Li;Songtao Guo","doi":"10.1109/TCC.2024.3452696","DOIUrl":"10.1109/TCC.2024.3452696","url":null,"abstract":"Vehicle speed prediction is crucial for intelligent transportation systems, promoting more reliable autonomous driving by accurately predicting future vehicle conditions. Due to variations in drivers’ driving styles and vehicle types, speed predictions for different target vehicles may significantly differ. Existing methods may not realize personalized vehicle speed prediction while protecting drivers’ data privacy. We propose a Federated learning framework with Personalized Aggregation Weights (FedPAW) to overcome these challenges. This method captures client-specific information by measuring the weighted mean squared error between the parameters of local models and global models. The server sends tailored aggregated models to clients instead of a single global model, without incurring additional computational and communication overhead for clients. To evaluate the effectiveness of FedPAW, we collected driving data in urban scenarios using the autonomous driving simulator CARLA, employing an LSTM-based Seq2Seq model with a multi-head attention mechanism to predict the future speed of target vehicles. The results demonstrate that our proposed FedPAW ranks lowest in prediction error within the time horizon of 10 seconds, with a 0.8% reduction in test MAE, compared to eleven representative benchmark baselines.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 4","pages":"1248-1259"},"PeriodicalIF":5.3,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142218871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Large-Scale Measurements and Optimizations on Latency in Edge Clouds 边缘云延迟的大规模测量和优化
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-08-30 DOI: 10.1109/TCC.2024.3452094
Heng Zhang;Shaoyuan Huang;Mengwei Xu;Deke Guo;Xiaofei Wang;Xin Wang;Victor C. M. Leung;Wenyu Wang
The emergence of next-generation latency-critical applications places strict requirements on network latency and stability. Edge cloud, an instantiated paradigm for edge computing, is gaining more and more attention due to its benefits of low latency. In this work, we make an in-depth investigation into the network QoS, especially end-to-end latency, at both spatial and temporal dimensions on a nationwide edge computing platform. Through the measurements, we collect a multi-variable large-scale real-world dataset on latency. We then quantify how the spatial-temporal factors affect the end-to-end latency, and verify the predictability of end-to-end latency. The results reveal the limitation of centralized clouds and illustrate how could edge clouds provide low and stable latency. Our results also point out that existing edge clouds merely increase the density of servers and ignore spatial-temporal factors, so they still suffer from high latency and fluctuations. Based on a quantified latency impact factor, we have proposed several optimization strategies for edge cloud latency and validated their effectiveness. We also propose a robust prototype edge cloud model based on lessons we learn from the measurement and evaluate its performance in the production environment. Evaluation result shows that edge clouds achieve 84.1% latency reduction with 0.5 ms latency fluctuation and 73.3% QoS improvement compared with the centralized clouds.
下一代延迟关键型应用程序的出现对网络延迟和稳定性提出了严格的要求。边缘云作为边缘计算的一种实例化范例,由于其低延迟的优点而受到越来越多的关注。在这项工作中,我们在一个全国性的边缘计算平台上,从空间和时间两个维度深入研究了网络QoS,特别是端到端延迟。通过测量,我们收集了一个关于延迟的多变量大规模真实数据集。然后,我们量化了时空因素如何影响端到端延迟,并验证了端到端延迟的可预测性。结果揭示了集中式云的局限性,并说明了边缘云如何提供低而稳定的延迟。我们的研究结果还指出,现有的边缘云只是增加了服务器的密度,而忽略了时空因素,因此它们仍然存在高延迟和波动。基于一个量化的延迟影响因子,我们提出了几种边缘云延迟优化策略,并验证了它们的有效性。我们还提出了一个鲁棒的原型边缘云模型,基于我们从测量中吸取的教训,并评估其在生产环境中的性能。评估结果表明,与集中式云相比,边缘云延迟降低84.1%,延迟波动0.5 ms, QoS提高73.3%。
{"title":"Large-Scale Measurements and Optimizations on Latency in Edge Clouds","authors":"Heng Zhang;Shaoyuan Huang;Mengwei Xu;Deke Guo;Xiaofei Wang;Xin Wang;Victor C. M. Leung;Wenyu Wang","doi":"10.1109/TCC.2024.3452094","DOIUrl":"10.1109/TCC.2024.3452094","url":null,"abstract":"The emergence of next-generation latency-critical applications places strict requirements on network latency and stability. Edge cloud, an instantiated paradigm for edge computing, is gaining more and more attention due to its benefits of low latency. In this work, we make an in-depth investigation into the network QoS, especially end-to-end latency, at both spatial and temporal dimensions on a nationwide edge computing platform. Through the measurements, we collect a multi-variable large-scale real-world dataset on latency. We then quantify how the spatial-temporal factors affect the end-to-end latency, and verify the predictability of end-to-end latency. The results reveal the limitation of centralized clouds and illustrate how could edge clouds provide low and stable latency. Our results also point out that existing edge clouds merely increase the density of servers and ignore spatial-temporal factors, so they still suffer from high latency and fluctuations. Based on a quantified latency impact factor, we have proposed several optimization strategies for edge cloud latency and validated their effectiveness. We also propose a robust prototype edge cloud model based on lessons we learn from the measurement and evaluate its performance in the production environment. Evaluation result shows that edge clouds achieve 84.1% latency reduction with 0.5 ms latency fluctuation and 73.3% QoS improvement compared with the centralized clouds.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 4","pages":"1218-1231"},"PeriodicalIF":5.3,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142218872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attribute-Based Hierarchical Keyword Auditing With Batch Fault Localization Assisted by Smart Contracts 基于属性的分层关键词审计与智能合约辅助下的批量故障定位
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-08-30 DOI: 10.1109/TCC.2024.3452324
Jingting Xue;Shuqin Luo;Fagen Li;Wenzheng Zhang;Liang Liu;Yu Zhou;Xiaojun Zhang
Keyword-based auditing (KA) provides a means for users to verify the integrity of only the outsourced data they are interested in. Existing KA schemes employ relation authentication labels to conduct targeted audits with keywords, which significantly improves the cost-effectiveness. However, such schemes typically support only a single-challenge scenario, which may not always be practical. To overcome this constraint, we introduce a hierarchical challenge mechanism grounded in user attributes. This mechanism leverages inequality and affiliation relationships to comply with a predefined tree structure for access policies. Incorporated during the challenge-response phase of the auditing model, it permits users to initiate cross-challenges. Expanding upon this hierarchical mechanism, we propose an attribute-based hierarchical keyword auditing scheme, abbreviated as $mathcal{AHKA}$. $mathcal{AHKA}$ combines searchable encryption to conduct cross-targeted audits and benefits from the hash collision mapping of Bloom filters to safeguard against keyword guessing attacks. Moreover, we design a fault localization algorithm based on a variant of the binary search technique. It locates in batch the faulty cloud servers and damaged data blocks after an audit failure. As an integral part of $mathcal{AHKA}$, the algorithm significantly enhances our scheme's practicability. Security analyses indicate that $mathcal{AHKA}$ can effectively withstand both forgery and replace attacks on audit proofs. The smart contract component ensures that our scheme's processes can be monitored and regulated. Experimental data corroborate that deploying $mathcal{AHKA}$ on the client side and on the blockchain is both efficient and feasible.
基于关键字的审计(KA)为用户提供了一种方法来验证他们感兴趣的外包数据的完整性。现有的KA方案采用关系身份验证标签对关键字进行有针对性的审计,大大提高了成本效益。然而,这样的方案通常只支持单一挑战场景,这可能并不总是实用的。为了克服这一限制,我们引入了基于用户属性的分层挑战机制。该机制利用不平等和从属关系来遵守访问策略的预定义树结构。在审计模型的质询-响应阶段合并,它允许用户发起交叉质询。在这个分层机制的基础上,我们提出了一个基于属性的分层关键字审计方案,缩写为$mathcal{AHKA}$。$mathcal{AHKA}$结合可搜索加密进行交叉目标审计,并受益于布隆过滤器的哈希冲突映射,以防止关键字猜测攻击。此外,我们设计了一种基于二叉搜索技术的故障定位算法。对审计失败后出现故障的云服务器和损坏的数据块进行批量定位。该算法作为$mathcal{AHKA}$的组成部分,大大提高了方案的实用性。安全分析表明,$mathcal{AHKA}$可以有效抵御审计证明的伪造和替换攻击。智能合约组件确保我们的方案流程可以被监控和监管。实验数据证实,在客户端和区块链上部署$mathcal{AHKA}$是有效和可行的。
{"title":"Attribute-Based Hierarchical Keyword Auditing With Batch Fault Localization Assisted by Smart Contracts","authors":"Jingting Xue;Shuqin Luo;Fagen Li;Wenzheng Zhang;Liang Liu;Yu Zhou;Xiaojun Zhang","doi":"10.1109/TCC.2024.3452324","DOIUrl":"10.1109/TCC.2024.3452324","url":null,"abstract":"Keyword-based auditing (KA) provides a means for users to verify the integrity of only the outsourced data they are interested in. Existing KA schemes employ relation authentication labels to conduct targeted audits with keywords, which significantly improves the cost-effectiveness. However, such schemes typically support only a single-challenge scenario, which may not always be practical. To overcome this constraint, we introduce a hierarchical challenge mechanism grounded in user attributes. This mechanism leverages inequality and affiliation relationships to comply with a predefined tree structure for access policies. Incorporated during the challenge-response phase of the auditing model, it permits users to initiate cross-challenges. Expanding upon this hierarchical mechanism, we propose an attribute-based hierarchical keyword auditing scheme, abbreviated as \u0000<inline-formula><tex-math>$mathcal{AHKA}$</tex-math></inline-formula>\u0000. \u0000<inline-formula><tex-math>$mathcal{AHKA}$</tex-math></inline-formula>\u0000 combines searchable encryption to conduct cross-targeted audits and benefits from the hash collision mapping of Bloom filters to safeguard against keyword guessing attacks. Moreover, we design a fault localization algorithm based on a variant of the binary search technique. It locates in batch the faulty cloud servers and damaged data blocks after an audit failure. As an integral part of \u0000<inline-formula><tex-math>$mathcal{AHKA}$</tex-math></inline-formula>\u0000, the algorithm significantly enhances our scheme's practicability. Security analyses indicate that \u0000<inline-formula><tex-math>$mathcal{AHKA}$</tex-math></inline-formula>\u0000 can effectively withstand both forgery and replace attacks on audit proofs. The smart contract component ensures that our scheme's processes can be monitored and regulated. Experimental data corroborate that deploying \u0000<inline-formula><tex-math>$mathcal{AHKA}$</tex-math></inline-formula>\u0000 on the client side and on the blockchain is both efficient and feasible.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 4","pages":"1232-1247"},"PeriodicalIF":5.3,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142218870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FUSIONIZE++: Improving Serverless Application Performance Using Dynamic Task Inlining and Infrastructure Optimization FUSIONIZE++:利用动态任务内嵌和基础设施优化提高无服务器应用程序性能
IF 5.3 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-08-28 DOI: 10.1109/TCC.2024.3451108
Trever Schirmer;Joel Scheuner;Tobias Pfandzelter;David Bermbach
The Function-as-a-Service (FaaS) execution model increases developer productivity by removing operational concerns such as managing hardware or software runtimes. Developers, however, still need to partition their applications into FaaS functions, which is error-prone and complex: Encapsulating only the smallest logical unit of an application as a FaaS function maximizes flexibility and reusability. Yet, it also leads to invocation overheads, additional cold starts, and may increase cost due to double billing during synchronous invocations. Conversely, deploying an entire application as a single FaaS function avoids these overheads but decreases flexibility. In this paper we present Fusionize, a framework that automates optimizing for this trade-off by automatically fusing application code into an optimized multi-function composition. Developers only need to write fine-grained application code following the serverless model, while Fusionize automatically fuses different parts of the application into FaaS functions, manages their interactions, and configures the underlying infrastructure. At runtime, it monitors application performance and adapts it to minimize request-response latency and costs. Real-world use cases show that Fusionize can improve the deployment artifacts of the application, reducing both median request-response latency and cost of an example IoT application by more than 35%.
功能即服务(FaaS)执行模型通过消除诸如管理硬件或软件运行时之类的操作问题来提高开发人员的工作效率。然而,开发人员仍然需要将他们的应用程序划分为FaaS功能,这是容易出错且复杂的:仅将应用程序的最小逻辑单元封装为FaaS功能,可以最大限度地提高灵活性和可重用性。然而,它也会导致调用开销、额外的冷启动,并可能由于同步调用期间的双重计费而增加成本。相反,将整个应用程序部署为单个FaaS功能可以避免这些开销,但会降低灵活性。在本文中,我们介绍了Fusionize,一个通过自动将应用程序代码融合到优化的多功能组合中来自动优化这种权衡的框架。开发人员只需要按照无服务器模型编写细粒度的应用程序代码,而Fusionize则自动将应用程序的不同部分融合到FaaS功能中,管理它们的交互,并配置底层基础设施。在运行时,它监视应用程序性能并对其进行调整,以最小化请求-响应延迟和成本。实际用例表明,Fusionize可以改善应用程序的部署工件,将示例物联网应用程序的请求响应延迟中值和成本降低35%以上。
{"title":"FUSIONIZE++: Improving Serverless Application Performance Using Dynamic Task Inlining and Infrastructure Optimization","authors":"Trever Schirmer;Joel Scheuner;Tobias Pfandzelter;David Bermbach","doi":"10.1109/TCC.2024.3451108","DOIUrl":"10.1109/TCC.2024.3451108","url":null,"abstract":"The Function-as-a-Service (FaaS) execution model increases developer productivity by removing operational concerns such as managing hardware or software runtimes. Developers, however, still need to partition their applications into FaaS functions, which is error-prone and complex: Encapsulating only the smallest logical unit of an application as a FaaS function maximizes flexibility and reusability. Yet, it also leads to invocation overheads, additional cold starts, and may increase cost due to double billing during synchronous invocations. Conversely, deploying an entire application as a single FaaS function avoids these overheads but decreases flexibility. In this paper we present \u0000<sc>Fusionize</small>\u0000, a framework that automates optimizing for this trade-off by automatically fusing application code into an optimized multi-function composition. Developers only need to write fine-grained application code following the serverless model, while \u0000<sc>Fusionize</small>\u0000 automatically fuses different parts of the application into FaaS functions, manages their interactions, and configures the underlying infrastructure. At runtime, it monitors application performance and adapts it to minimize request-response latency and costs. Real-world use cases show that \u0000<sc>Fusionize</small>\u0000 can improve the deployment artifacts of the application, reducing both median request-response latency and cost of an example IoT application by more than 35%.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 4","pages":"1172-1185"},"PeriodicalIF":5.3,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142218874","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Cloud Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1