首页 > 最新文献

Journal of Parallel and Distributed Computing最新文献

英文 中文
PerfTop: Towards performance prediction of distributed learning over general topology PerfTop:在一般拓扑结构上实现分布式学习的性能预测
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-24 DOI: 10.1016/j.jpdc.2024.104922
Changzhi Yan, Zehan Zhu, Youcheng Niu, Cong Wang, Cheng Zhuo, Jinming Xu

Distributed learning with multiple GPUs has been widely adopted to accelerate the training process of large-scale deep neural networks. However, misconfiguration of the GPU clusters with various communication primitives and topologies could potentially diminish the gains in parallel computation and lead to significant degradation in training efficiency. Predicting the performance of distributed learning enables service providers to identify potential bottlenecks beforehand. In this work, we propose a Performance prediction framework over General Topologies, called PerfTop, for accurate estimation of per-iteration execution time. The main strategy is to integrate computation time prediction with an analytical model to map the nonlinearity in communication and fine-grained computation-communication patterns. This enables accurate prediction of a variety of neural network models over general topologies, such as tree, hierarchical, and exponential. Our extensive experiments show that PerfTop outperforms existing methods in estimating both computation and communication time, particularly for communication, surpassing the existing methods by over 45%. Meanwhile, it achieves an accuracy of above 85% in predicting the execution time over general topologies compared to simple topologies such as star and ring from the previous works.

利用多个 GPU 进行分布式学习已被广泛采用,以加速大规模深度神经网络的训练过程。然而,利用各种通信原语和拓扑结构对 GPU 集群进行错误配置,可能会降低并行计算的收益,导致训练效率显著下降。预测分布式学习的性能能让服务提供商提前发现潜在瓶颈。在这项工作中,我们提出了一个名为 PerfTop 的通用拓扑性能预测框架,用于准确估算每次迭代的执行时间。主要策略是将计算时间预测与分析模型相结合,以映射通信中的非线性和细粒度计算-通信模式。这样就能准确预测各种神经网络模型的一般拓扑结构,如树型、分层型和指数型。我们的大量实验表明,PerfTop 在估算计算和通信时间方面都优于现有方法,尤其是在通信时间方面,超过现有方法 45% 以上。同时,在预测一般拓扑结构的执行时间时,它的准确率达到了 85% 以上,而之前的研究只预测了星形和环形等简单拓扑结构的执行时间。
{"title":"PerfTop: Towards performance prediction of distributed learning over general topology","authors":"Changzhi Yan,&nbsp;Zehan Zhu,&nbsp;Youcheng Niu,&nbsp;Cong Wang,&nbsp;Cheng Zhuo,&nbsp;Jinming Xu","doi":"10.1016/j.jpdc.2024.104922","DOIUrl":"10.1016/j.jpdc.2024.104922","url":null,"abstract":"<div><p>Distributed learning with multiple GPUs has been widely adopted to accelerate the training process of large-scale deep neural networks. However, misconfiguration of the GPU clusters with various communication primitives and topologies could potentially diminish the gains in parallel computation and lead to significant degradation in training efficiency. Predicting the performance of distributed learning enables service providers to identify potential bottlenecks beforehand. In this work, we propose a <u>Perf</u>ormance prediction framework over General <u>Top</u>ologies, called PerfTop, for accurate estimation of per-iteration execution time. The main strategy is to integrate computation time prediction with an analytical model to map the nonlinearity in communication and fine-grained computation-communication patterns. This enables accurate prediction of a variety of neural network models over general topologies, such as tree, hierarchical, and exponential. Our extensive experiments show that PerfTop outperforms existing methods in estimating both computation and communication time, particularly for communication, surpassing the existing methods by over 45%. Meanwhile, it achieves an accuracy of above 85% in predicting the execution time over general topologies compared to simple topologies such as star and ring from the previous works.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141139181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Local outlier factor for anomaly detection in HPCC systems 用于 HPCC 系统异常检测的局部离群因子
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-23 DOI: 10.1016/j.jpdc.2024.104923
Arya Adesh , Shobha G , Jyoti Shetty , Lili Xu

Local Outlier Factor (LOF) is an unsupervised anomaly detection algorithm that finds anomalies by assessing the local density of a data point relative to its neighborhood. Anomaly detection is the process of finding anomalies in datasets. Anomalies in real-time datasets may indicate critical events like bank frauds, data compromise, network threats, etc. This paper deals with the implementation of the LOF algorithm in the HPCC Systems platform, which is an open-source distributed computing platform for big data analytics. Improved LOF is also proposed which efficiently detects anomalies in datasets rich in duplicates. The impact of varying hyperparameters on the performance of LOF is examined in HPCC Systems. This paper examines the performance of LOF with other algorithms like COF, LoOP, and kNN over several datasets in the HPCC Systems. Additionally, the efficacy of LOF is evaluated across big-data frameworks such as Spark, Hadoop, and HPCC Systems, by comparing their runtime performances.

局部离群因子(LOF)是一种无监督异常检测算法,它通过评估数据点相对于其邻域的局部密度来发现异常。异常检测是在数据集中发现异常的过程。实时数据集中的异常可能预示着银行欺诈、数据泄露、网络威胁等重大事件。本文论述了 LOF 算法在 HPCC 系统平台上的实现,该平台是用于大数据分析的开源分布式计算平台。本文还提出了改进的 LOF 算法,它能有效地检测出重复数据集中的异常情况。在 HPCC 系统中,研究了不同超参数对 LOF 性能的影响。本文通过 HPCC 系统中的多个数据集,检验了 LOF 与 COF、LoOP 和 kNN 等其他算法的性能。此外,通过比较 Spark、Hadoop 和 HPCC 系统等大数据框架的运行性能,评估了 LOF 的功效。
{"title":"Local outlier factor for anomaly detection in HPCC systems","authors":"Arya Adesh ,&nbsp;Shobha G ,&nbsp;Jyoti Shetty ,&nbsp;Lili Xu","doi":"10.1016/j.jpdc.2024.104923","DOIUrl":"10.1016/j.jpdc.2024.104923","url":null,"abstract":"<div><p>Local Outlier Factor (LOF) is an unsupervised anomaly detection algorithm that finds anomalies by assessing the local density of a data point relative to its neighborhood. Anomaly detection is the process of finding anomalies in datasets. Anomalies in real-time datasets may indicate critical events like bank frauds, data compromise, network threats, etc. This paper deals with the implementation of the LOF algorithm in the HPCC Systems platform, which is an open-source distributed computing platform for big data analytics. Improved LOF is also proposed which efficiently detects anomalies in datasets rich in duplicates. The impact of varying hyperparameters on the performance of LOF is examined in HPCC Systems. This paper examines the performance of LOF with other algorithms like COF, LoOP, and kNN over several datasets in the HPCC Systems. Additionally, the efficacy of LOF is evaluated across big-data frameworks such as Spark, Hadoop, and HPCC Systems, by comparing their runtime performances.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141140257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GraMeR: Graph Meta Reinforcement learning for multi-objective influence maximization GraMeR: 面向多目标影响力最大化的图元强化学习
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-23 DOI: 10.1016/j.jpdc.2024.104900
Sai Munikoti , Balasubramaniam Natarajan , Mahantesh Halappanavar

Influence maximization (IM) is a combinatorial problem of identifying a subset of seed nodes in a network (graph), which when activated, provide a maximal spread of influence in the network for a given diffusion model and a budget for seed set size. IM has numerous applications such as viral marketing, epidemic control, sensor placement and other network-related tasks. However, its practical uses are limited due to the computational complexity of current algorithms. Recently, deep reinforcement learning has been leveraged to solve IM in order to ease the computational burden. However, there are serious limitations in current approaches, including narrow IM formulation that only consider influence via spread and ignore self activation, low scalability to large graphs, and lack of generalizability across graph families leading to a large running time for every test network. In this work, we address these limitations through a unique approach that involves: (1) Formulating a generic IM problem as a Markov decision process that handles both intrinsic and influence activations; (2) incorporating generalizability via meta-learning across graph families. There are previous works that combine deep reinforcement learning with graph neural network but this work solves a more realistic IM problem and incorporates generalizability across graphs via meta reinforcement learning. Extensive experiments are carried out in various standard networks to validate performance of the proposed Graph Meta Reinforcement learning (GraMeR) framework. The results indicate that GraMeR is multiple orders faster and generic than conventional approaches when applied on small to medium scale graphs.

影响最大化(IM)是一个组合问题,即在网络(图)中确定一个种子节点子集,当激活该子集时,在给定的扩散模型和种子集大小预算下,该子集可在网络中提供最大的影响传播。IM 有许多应用,如病毒营销、流行病控制、传感器安置和其他网络相关任务。然而,由于当前算法的计算复杂性,其实际应用受到了限制。最近,人们利用深度强化学习来解决 IM 问题,以减轻计算负担。然而,目前的方法存在严重的局限性,包括只考虑通过传播产生影响而忽略自激活的狭隘 IM 表述、对大型图的可扩展性低、缺乏跨图族的泛化能力,导致每个测试网络的运行时间都很长。在这项研究中,我们采用了一种独特的方法来解决这些局限性,其中包括:(1)将一般的 IM 问题表述为一个马尔可夫决策过程,该过程可同时处理内在激活和影响激活;(2)通过元学习在图族间实现通用性。之前有研究将深度强化学习与图神经网络相结合,但本研究解决的是一个更现实的 IM 问题,并通过元强化学习实现了跨图的通用性。我们在各种标准网络中进行了广泛的实验,以验证所提出的图元强化学习(GraMeR)框架的性能。结果表明,与传统方法相比,GraMeR 在中小型图上的应用速度和通用性要快上数倍。
{"title":"GraMeR: Graph Meta Reinforcement learning for multi-objective influence maximization","authors":"Sai Munikoti ,&nbsp;Balasubramaniam Natarajan ,&nbsp;Mahantesh Halappanavar","doi":"10.1016/j.jpdc.2024.104900","DOIUrl":"https://doi.org/10.1016/j.jpdc.2024.104900","url":null,"abstract":"<div><p>Influence maximization (IM) is a combinatorial problem of identifying a subset of seed nodes in a network (graph), which when activated, provide a maximal spread of influence in the network for a given diffusion model and a budget for seed set size. IM has numerous applications such as viral marketing, epidemic control, sensor placement and other network-related tasks. However, its practical uses are limited due to the computational complexity of current algorithms. Recently, deep reinforcement learning has been leveraged to solve IM in order to ease the computational burden. However, there are serious limitations in current approaches, including narrow IM formulation that only consider influence via spread and ignore self activation, low scalability to large graphs, and lack of generalizability across graph families leading to a large running time for every test network. In this work, we address these limitations through a unique approach that involves: (1) Formulating a generic IM problem as a Markov decision process that handles both intrinsic and influence activations; (2) incorporating generalizability via meta-learning across graph families. There are previous works that combine deep reinforcement learning with graph neural network but this work solves a more realistic IM problem and incorporates generalizability across graphs via meta reinforcement learning. Extensive experiments are carried out in various standard networks to validate performance of the proposed Graph Meta Reinforcement learning (GraMeR) framework. The results indicate that GraMeR is multiple orders faster and generic than conventional approaches when applied on small to medium scale graphs.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141423534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Large-scale and cooperative graybox parallel optimization on the supercomputer Fugaku 超级计算机 "Fugaku "上的大规模协同灰箱并行优化
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-22 DOI: 10.1016/j.jpdc.2024.104921
Lorenzo Canonne , Bilel Derbel , Miwako Tsuji , Mitsuhisa Sato

We design, develop and analyze parallel variants of a state-of-the-art graybox optimization algorithm, namely Drils (Deterministic Recombination and Iterated Local Search), for attacking large-scale pseudo-boolean optimization problems on top of the large-scale computing facilities offered by the supercomputer Fugaku. We first adopt a Master/Worker design coupled with a fully distributed Island-based model, ending up with a number of hybrid OpenMP/MPI implementations of high-level parallel Drils versions. We show that such a design, although effective, can be substantially improved by enabling a more focused iteration-level cooperation mechanism between the core graybox components of the original serial Drils algorithm. Extensive experiments are conducted in order to provide a systematic analysis of the impact of the designed parallel algorithms on search behavior, and their ability to compute high-quality solutions using increasing number of CPU-cores. Results using up to 1024×12-cores NUMA nodes, and NK-landscapes with up to 10,000 binary variables are reported, providing evidence on the relative strength of the designed hybrid cooperative graybox parallel search.

我们设计、开发并分析了一种最先进的灰盒优化算法的并行变体,即 Drils(确定性重组和迭代局部搜索),用于在超级计算机富加库提供的大规模计算设施之上解决大规模伪布尔优化问题。我们首先采用了 Master/Worker 设计和基于岛的完全分布式模型,最终得到了一些高级并行 Drils 版本的 OpenMP/MPI 混合实现。我们的研究表明,这种设计虽然有效,但可以通过在原始串行 Drils 算法的核心灰盒组件之间建立更集中的迭代级合作机制来大幅改进。我们进行了广泛的实验,以便系统分析所设计的并行算法对搜索行为的影响,以及使用越来越多的 CPU 核心计算高质量解决方案的能力。报告了使用多达 1024×12 核 NUMA 节点和多达 10,000 个二进制变量的 NK-landscapes 的结果,为所设计的混合合作灰箱并行搜索的相对优势提供了证据。
{"title":"Large-scale and cooperative graybox parallel optimization on the supercomputer Fugaku","authors":"Lorenzo Canonne ,&nbsp;Bilel Derbel ,&nbsp;Miwako Tsuji ,&nbsp;Mitsuhisa Sato","doi":"10.1016/j.jpdc.2024.104921","DOIUrl":"https://doi.org/10.1016/j.jpdc.2024.104921","url":null,"abstract":"<div><p>We design, develop and analyze parallel variants of a state-of-the-art graybox optimization algorithm, namely <span>Drils</span> (Deterministic Recombination and Iterated Local Search), for attacking large-scale pseudo-boolean optimization problems on top of the large-scale computing facilities offered by the supercomputer Fugaku. We first adopt a Master/Worker design coupled with a fully distributed Island-based model, ending up with a number of hybrid OpenMP/MPI implementations of high-level parallel <span>Drils</span> versions. We show that such a design, although effective, can be substantially improved by enabling a more focused iteration-level cooperation mechanism between the core graybox components of the original serial <span>Drils</span> algorithm. Extensive experiments are conducted in order to provide a systematic analysis of the impact of the designed parallel algorithms on search behavior, and their ability to compute high-quality solutions using increasing number of CPU-cores. Results using up to 1024×12-cores NUMA nodes, and NK-landscapes with up to <span><math><mn>10</mn><mo>,</mo><mn>000</mn></math></span> binary variables are reported, providing evidence on the relative strength of the designed hybrid cooperative graybox parallel search.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141090688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HBPB, applying reuse distance to improve cache efficiency proactively HBPB,应用重用距离主动提高高速缓存效率
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-20 DOI: 10.1016/j.jpdc.2024.104919
Arthur M. Krause, Paulo C. Santos, Arthur F. Lorenzon, Philippe O.A. Navaux

Cache memories play a significant role in the performance, area, and energy consumption of modern processors, and this impact is expected to grow as on-die memories become larger. While caches are highly effective for cache-friendly access patterns, they introduce unnecessary delays and energy wastage when they fail to serve the required data. Hence, cache bypassing techniques have been proposed to optimize the latency of cache-unfriendly memory accesses. In this scenario, we discuss HBPB, a history-based preemptive bypassing technique that accelerates cache-unfriendly access through the reduced latency of bypassing the caches. By extensively evaluating different real-world applications and hardware cache configurations, we show that HBPB yields energy reductions of up to 75% and performance improvements of up to 50% compared to a version that does not apply cache bypassing. More importantly, we demonstrate that HBPB does not affect the performance of applications with cache-friendly access patterns.

高速缓冲存储器在现代处理器的性能、面积和能耗方面发挥着重要作用,而且随着芯片上存储器的增大,预计这种影响还会越来越大。虽然高速缓存对高速缓存友好的访问模式非常有效,但当高速缓存无法提供所需数据时,就会带来不必要的延迟和能源浪费。因此,有人提出了高速缓存旁路技术,以优化不适合高速缓存的内存访问延迟。在本方案中,我们讨论了 HBPB,这是一种基于历史记录的抢先绕过技术,可通过缩短绕过高速缓存的延迟来加速高速缓存不友好访问。通过广泛评估不同的实际应用和硬件缓存配置,我们发现,与不应用缓存旁路的版本相比,HBPB 可减少高达 75% 的能耗,提高高达 50% 的性能。更重要的是,我们证明 HBPB 不会影响具有缓存友好访问模式的应用程序的性能。
{"title":"HBPB, applying reuse distance to improve cache efficiency proactively","authors":"Arthur M. Krause,&nbsp;Paulo C. Santos,&nbsp;Arthur F. Lorenzon,&nbsp;Philippe O.A. Navaux","doi":"10.1016/j.jpdc.2024.104919","DOIUrl":"https://doi.org/10.1016/j.jpdc.2024.104919","url":null,"abstract":"<div><p>Cache memories play a significant role in the performance, area, and energy consumption of modern processors, and this impact is expected to grow as on-die memories become larger. While caches are highly effective for cache-friendly access patterns, they introduce unnecessary delays and energy wastage when they fail to serve the required data. Hence, cache bypassing techniques have been proposed to optimize the latency of cache-unfriendly memory accesses. In this scenario, we discuss <em>HBPB</em>, a history-based preemptive bypassing technique that accelerates cache-unfriendly access through the reduced latency of bypassing the caches. By extensively evaluating different real-world applications and hardware cache configurations, we show that <em>HBPB</em> yields energy reductions of up to 75% and performance improvements of up to 50% compared to a version that does not apply cache bypassing. More importantly, we demonstrate that <em>HBPB</em> does not affect the performance of applications with cache-friendly access patterns.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141078565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Balancing privacy and performance in federated learning: A systematic literature review on methods and metrics 联盟学习中的隐私与性能平衡:关于方法和指标的系统性文献综述
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-17 DOI: 10.1016/j.jpdc.2024.104918
Samaneh Mohammadi , Ali Balador , Sima Sinaei , Francesco Flammini

Federated learning (FL) as a novel paradigm in Artificial Intelligence (AI), ensures enhanced privacy by eliminating data centralization and brings learning directly to the edge of the user's device. Nevertheless, new privacy issues have been raised particularly during training and the exchange of parameters between servers and clients. While several privacy-preserving FL solutions have been developed to mitigate potential breaches in FL architectures, their integration poses its own set of challenges. Incorporating these privacy-preserving mechanisms into FL at the edge computing level can increase both communication and computational overheads, which may, in turn, compromise data utility and learning performance metrics. This paper provides a systematic literature review on essential methods and metrics to support the most appropriate trade-offs between FL privacy and other performance-related application requirements such as accuracy, loss, convergence time, utility, communication, and computation overhead. We aim to provide an extensive overview of recent privacy-preserving mechanisms in FL used across various applications, placing a particular focus on quantitative privacy assessment approaches in FL and the necessity of achieving a balance between privacy and the other requirements of real-world FL applications. This review collects, classifies, and discusses relevant papers in a structured manner, emphasizing challenges, open issues, and promising research directions.

联合学习(FL)作为人工智能(AI)领域的一种新模式,通过消除数据集中并将学习直接带到用户设备的边缘,确保了更高的隐私性。然而,新的隐私问题也随之而来,尤其是在训练以及服务器和客户端之间交换参数的过程中。虽然已经开发了几种保护隐私的 FL 解决方案来减少 FL 架构中潜在的漏洞,但它们的整合也带来了一系列挑战。在边缘计算层面将这些隐私保护机制纳入 FL 可能会增加通信和计算开销,进而可能会影响数据效用和学习性能指标。本文对基本方法和指标进行了系统的文献综述,以支持在 FL 隐私和其他性能相关应用要求(如准确性、损失、收敛时间、效用、通信和计算开销)之间进行最适当的权衡。我们的目标是广泛概述最近在各种应用中使用的 FL 隐私保护机制,特别关注 FL 中的定量隐私评估方法,以及在隐私和真实世界 FL 应用的其他要求之间实现平衡的必要性。本综述以结构化的方式对相关论文进行了收集、分类和讨论,强调了挑战、开放性问题和有前景的研究方向。
{"title":"Balancing privacy and performance in federated learning: A systematic literature review on methods and metrics","authors":"Samaneh Mohammadi ,&nbsp;Ali Balador ,&nbsp;Sima Sinaei ,&nbsp;Francesco Flammini","doi":"10.1016/j.jpdc.2024.104918","DOIUrl":"10.1016/j.jpdc.2024.104918","url":null,"abstract":"<div><p>Federated learning (FL) as a novel paradigm in Artificial Intelligence (AI), ensures enhanced privacy by eliminating data centralization and brings learning directly to the edge of the user's device. Nevertheless, new privacy issues have been raised particularly during training and the exchange of parameters between servers and clients. While several privacy-preserving FL solutions have been developed to mitigate potential breaches in FL architectures, their integration poses its own set of challenges. Incorporating these privacy-preserving mechanisms into FL at the edge computing level can increase both communication and computational overheads, which may, in turn, compromise data utility and learning performance metrics. This paper provides a systematic literature review on essential methods and metrics to support the most appropriate trade-offs between FL privacy and other performance-related application requirements such as accuracy, loss, convergence time, utility, communication, and computation overhead. We aim to provide an extensive overview of recent privacy-preserving mechanisms in FL used across various applications, placing a particular focus on quantitative privacy assessment approaches in FL and the necessity of achieving a balance between privacy and the other requirements of real-world FL applications. This review collects, classifies, and discusses relevant papers in a structured manner, emphasizing challenges, open issues, and promising research directions.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0743731524000820/pdfft?md5=2ee3078ecc6441a5efe38d3a7c047d80&pid=1-s2.0-S0743731524000820-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141033667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cost-aware quantum-inspired genetic algorithm for workflow scheduling in hybrid clouds 用于混合云中工作流调度的成本感知量子启发遗传算法
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-17 DOI: 10.1016/j.jpdc.2024.104920
Mehboob Hussain , Lian-Fu Wei , Amir Rehman , Muqadar Ali , Syed Muhammad Waqas , Fakhar Abbas

Cloud computing delivers a desirable environment for users to run their different kinds of applications in a cloud. Numerous of these applications (tasks), such as bioinformatics, astronomy, biodiversity, and image analysis, are deadline-sensitive. Such tasks must be properly allocated to virtual machines (VMs) to avoid deadline violations, and they should reduce their execution time and cost. Due to the contradictory environment, minimizing the application task's completion time and execution cost is extremely difficult. Thus, we propose a Cost-aware Quantum-inspired Genetic Algorithm (CQGA) to minimize the execution time and cost by meeting the deadline constraints. CQGA is motivated by quantum computing and genetic algorithm. It combines quantum operators (measure, interference, and rotation) with genetic operators (selection, crossover, and mutation). Quantum operators are used for better population diversity, quick convergence, time-saving, and robustness. Genetic operators help to produce new individuals, have good fitness values for individuals, and play a significant role in preserving the evolution quality of the population. In addition, CQGA used a quantum bit as a probabilistic representation because it has higher population diversity attributes than other representations. The simulation outcome exhibits that the proposed algorithm can obtain outstanding convergence performance and reduced maximum cost than benchmark algorithms.

云计算为用户在云中运行不同类型的应用程序提供了理想的环境。生物信息学、天文学、生物多样性和图像分析等许多应用(任务)对截止日期非常敏感。必须将这些任务适当地分配到虚拟机(VM)上,以避免违反截止日期,并减少其执行时间和成本。由于环境的矛盾,最大限度地减少应用任务的完成时间和执行成本极其困难。因此,我们提出了一种成本感知量子启发遗传算法(CQGA),在满足截止日期约束的前提下最大限度地减少执行时间和成本。CQGA 的灵感来自量子计算和遗传算法。它结合了量子算子(测量、干涉和旋转)和遗传算子(选择、交叉和突变)。量子算子用于提高种群多样性、快速收敛、节省时间和鲁棒性。遗传算子有助于产生新个体,为个体提供良好的适应度值,并在保持种群进化质量方面发挥重要作用。此外,CQGA 使用量子比特作为概率表示,因为与其他表示相比,量子比特具有更高的种群多样性属性。仿真结果表明,与基准算法相比,所提出的算法能获得出色的收敛性能,并降低了最大成本。
{"title":"Cost-aware quantum-inspired genetic algorithm for workflow scheduling in hybrid clouds","authors":"Mehboob Hussain ,&nbsp;Lian-Fu Wei ,&nbsp;Amir Rehman ,&nbsp;Muqadar Ali ,&nbsp;Syed Muhammad Waqas ,&nbsp;Fakhar Abbas","doi":"10.1016/j.jpdc.2024.104920","DOIUrl":"10.1016/j.jpdc.2024.104920","url":null,"abstract":"<div><p>Cloud computing delivers a desirable environment for users to run their different kinds of applications in a cloud. Numerous of these applications (tasks), such as bioinformatics, astronomy, biodiversity, and image analysis, are deadline-sensitive. Such tasks must be properly allocated to virtual machines (VMs) to avoid deadline violations, and they should reduce their execution time and cost. Due to the contradictory environment, minimizing the application task's completion time and execution cost is extremely difficult. Thus, we propose a Cost-aware Quantum-inspired Genetic Algorithm (CQGA) to minimize the execution time and cost by meeting the deadline constraints. CQGA is motivated by quantum computing and genetic algorithm. It combines quantum operators (measure, interference, and rotation) with genetic operators (selection, crossover, and mutation). Quantum operators are used for better population diversity, quick convergence, time-saving, and robustness. Genetic operators help to produce new individuals, have good fitness values for individuals, and play a significant role in preserving the evolution quality of the population. In addition, CQGA used a quantum bit as a probabilistic representation because it has higher population diversity attributes than other representations. The simulation outcome exhibits that the proposed algorithm can obtain outstanding convergence performance and reduced maximum cost than benchmark algorithms.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141047734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CCFTL: A novel continuity compressed page-level flash address mapping method for SSDs CCFTL:适用于固态硬盘的新型连续性压缩页面级闪存地址映射方法
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-15 DOI: 10.1016/j.jpdc.2024.104917
Liangkuan Su , Mingwei Lin , Jianpeng Zhang , Yubiao Pan

Given the distinctive characteristics of flash-based solid-state drives (SSDs), such as out-of-place update scheme, as compared to traditional block storage devices, a flash translation layer (FTL) has been introduced to hide these features. In the FTL, there is an address translation module that implements the conversion from logical addresses to physical addresses. However, existing address mapping algorithms currently fail to fully exploit the mapping information generated by large I/O requests. First, based on this observation, we propose a novel continuity compressed page-level flash address mapping method (CCFTL). This method effectively compresses the mapping relationship between consecutive logical addresses and physical addresses, enabling the storage of more mapping information within the same mapping cache size. Next, we introduce two-level LRU linked list to mitigate the issue of compressed mapping entry splitting that arises from handling write requests. Finally, our experiments show that CCFTL reduced average response times by 52.67%, 16.81%, and 12.71% compared to DFTL, TPFTL, and MFTL, respectively. As the mapping cache size decreases from 2 MB to 1 MB, then further decreases to 256 KB, 128 KB, and eventually down to 64 KB, CCFTL experiences an average decline ratio of less than 3% in average response time, while the other three algorithms show an average decline ratio of 9.51%.

与传统的块存储设备相比,基于闪存的固态硬盘(SSD)具有不同的特性,如非就地更新方案,因此引入了闪存转换层(FTL)来隐藏这些特性。在 FTL 中,有一个地址转换模块可以实现从逻辑地址到物理地址的转换。然而,现有的地址映射算法目前无法充分利用大型 I/O 请求产生的映射信息。首先,基于这一观点,我们提出了一种新颖的连续性压缩页面级闪存地址映射方法(CCFTL)。这种方法能有效压缩连续逻辑地址和物理地址之间的映射关系,从而在相同的映射缓存大小内存储更多的映射信息。接下来,我们引入了两级 LRU 链接列表,以缓解处理写入请求时出现的压缩映射条目分割问题。最后,我们的实验表明,与 DFTL、TPFTL 和 MFTL 相比,CCFTL 的平均响应时间分别缩短了 52.67%、16.81% 和 12.71%。随着映射缓存大小从 2 MB 减小到 1 MB,然后进一步减小到 256 KB、128 KB,最终减小到 64 KB,CCFTL 的平均响应时间平均下降率不到 3%,而其他三种算法的平均下降率为 9.51%。
{"title":"CCFTL: A novel continuity compressed page-level flash address mapping method for SSDs","authors":"Liangkuan Su ,&nbsp;Mingwei Lin ,&nbsp;Jianpeng Zhang ,&nbsp;Yubiao Pan","doi":"10.1016/j.jpdc.2024.104917","DOIUrl":"10.1016/j.jpdc.2024.104917","url":null,"abstract":"<div><p>Given the distinctive characteristics of flash-based solid-state drives (SSDs), such as out-of-place update scheme, as compared to traditional block storage devices, a flash translation layer (FTL) has been introduced to hide these features. In the FTL, there is an address translation module that implements the conversion from logical addresses to physical addresses. However, existing address mapping algorithms currently fail to fully exploit the mapping information generated by large I/O requests. First, based on this observation, we propose a novel continuity compressed page-level flash address mapping method (CCFTL). This method effectively compresses the mapping relationship between consecutive logical addresses and physical addresses, enabling the storage of more mapping information within the same mapping cache size. Next, we introduce two-level LRU linked list to mitigate the issue of compressed mapping entry splitting that arises from handling write requests. Finally, our experiments show that CCFTL reduced average response times by 52.67%, 16.81%, and 12.71% compared to DFTL, TPFTL, and MFTL, respectively. As the mapping cache size decreases from 2 MB to 1 MB, then further decreases to 256 KB, 128 KB, and eventually down to 64 KB, CCFTL experiences an average decline ratio of less than 3% in average response time, while the other three algorithms show an average decline ratio of 9.51%.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141046274","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Federated variational generative learning for heterogeneous data in distributed environments 针对分布式环境中异构数据的联合变式生成学习
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-14 DOI: 10.1016/j.jpdc.2024.104916
Wei Xie, Runqun Xiong, Jinghui Zhang, Jiahui Jin, Junzhou Luo

Distributedly training models across diverse clients with heterogeneous data samples can significantly impact the convergence of federated learning. Various novel federated learning methods address these challenges but often require significant communication resources and local computational capacity, leading to reduced global inference accuracy in scenarios with imbalanced label data distribution and quantity skew. To tackle these challenges, we propose FedVGL, a Federated Variational Generative Learning method that directly trains a local generative model to learn the distribution of local features and improve global target model inference accuracy during aggregation, particularly under conditions of severe data heterogeneity. FedVGL facilitates distributed learning by sharing generators and latent vectors with the global server, aiding in global target model training from mapping local data distribution to the variational latent space for feature reconstruction. Additionally, FedVGL implements anonymization and encryption techniques to bolster privacy during generative model transmission and aggregation. In comparison to vanilla federated learning, FedVGL minimizes communication overhead, demonstrating superior accuracy even with minimal communication rounds. It effectively mitigates model drift in scenarios with heterogeneous data, delivering improved target model training outcomes. Empirical results establish FedVGL's superiority over baseline federated learning methods under severe label imbalance and data skew condition. In a Label-based Dirichlet Distribution setting with α=0.01 and 10 clients using the MNIST dataset, FedVGL achieved an exceptional accuracy over 97% with the VGG-9 target model.

在具有异构数据样本的不同客户端上分布式训练模型,会严重影响联合学习的收敛性。各种新颖的联合学习方法都能应对这些挑战,但往往需要大量通信资源和本地计算能力,导致在标签数据分布不平衡和数量倾斜的情况下,全局推断的准确性降低。为了应对这些挑战,我们提出了 FedVGL,这是一种联合变异生成学习方法,它能直接训练局部生成模型,以学习局部特征的分布,并在聚合过程中提高全局目标模型推断的准确性,尤其是在数据异构严重的情况下。FedVGL 通过与全局服务器共享生成器和潜向量来促进分布式学习,通过将本地数据分布映射到用于特征重构的变异潜空间来帮助全局目标模型训练。此外,FedVGL 还采用了匿名和加密技术,以在生成模型传输和聚合过程中保护隐私。与传统的联合学习相比,FedVGL 最大限度地减少了通信开销,即使在通信轮数极少的情况下也能显示出卓越的准确性。它能有效缓解异构数据场景中的模型漂移,从而改善目标模型的训练结果。实证结果表明,在严重的标签不平衡和数据倾斜条件下,FedVGL 比基线联合学习方法更具优势。在基于标签的 Dirichlet 分布设置(α=0.01)和 10 个客户端使用 MNIST 数据集的情况下,FedVGL 的 VGG-9 目标模型的准确率超过了 97%。
{"title":"Federated variational generative learning for heterogeneous data in distributed environments","authors":"Wei Xie,&nbsp;Runqun Xiong,&nbsp;Jinghui Zhang,&nbsp;Jiahui Jin,&nbsp;Junzhou Luo","doi":"10.1016/j.jpdc.2024.104916","DOIUrl":"10.1016/j.jpdc.2024.104916","url":null,"abstract":"<div><p>Distributedly training models across diverse clients with heterogeneous data samples can significantly impact the convergence of federated learning. Various novel federated learning methods address these challenges but often require significant communication resources and local computational capacity, leading to reduced global inference accuracy in scenarios with imbalanced label data distribution and quantity skew. To tackle these challenges, we propose FedVGL, a Federated Variational Generative Learning method that directly trains a local generative model to learn the distribution of local features and improve global target model inference accuracy during aggregation, particularly under conditions of severe data heterogeneity. FedVGL facilitates distributed learning by sharing generators and latent vectors with the global server, aiding in global target model training from mapping local data distribution to the variational latent space for feature reconstruction. Additionally, FedVGL implements anonymization and encryption techniques to bolster privacy during generative model transmission and aggregation. In comparison to vanilla federated learning, FedVGL minimizes communication overhead, demonstrating superior accuracy even with minimal communication rounds. It effectively mitigates model drift in scenarios with heterogeneous data, delivering improved target model training outcomes. Empirical results establish FedVGL's superiority over baseline federated learning methods under severe label imbalance and data skew condition. In a Label-based Dirichlet Distribution setting with <em>α</em>=0.01 and 10 clients using the MNIST dataset, FedVGL achieved an exceptional accuracy over 97% with the VGG-9 target model.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141036748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Energy-efficient triple modular redundancy scheduling on heterogeneous multi-core real-time systems 异构多核实时系统上的高能效三模块冗余调度
IF 3.8 3区 计算机科学 Q1 Mathematics Pub Date : 2024-05-13 DOI: 10.1016/j.jpdc.2024.104915
Hongzhi Xu , Binlian Zhang , Chen Pan , Keqin Li

Triple modular redundancy (TMR) fault tolerance mechanism can provide almost perfect fault-masking, which has the great potential to enhance the reliability of real-time systems. However, multiple copies of a task are executed concurrently, which will lead to a sharp increase in system energy consumption. In this work, the problem of parallel applications using TMR on heterogeneous multi-core platforms to minimize energy consumption is studied. First, the heterogeneous earliest finish time algorithm is improved, and then according to the given application's deadline constraints and reliability requirements, an algorithm to extend the execution time of the copies is designed. Secondly, based on the properties of TMR, an algorithm for minimizing the execution overhead of the third copy (MEOTC) is designed. Finally, considering the actual situation of task execution, an online energy management (OEM) method is proposed. The proposed algorithms were compared with the state-of-the-art AFTSA algorithm, and the results show significant differences in energy consumption. Specifically, for light fault detection, the energy consumption of the MEOTC and OEM algorithms was found to be 80% and 72% respectively, compared with AFTSA. For heavy fault detection, the energy consumption of MEOTC and OEM was measured at 61% and 55% respectively, compared with AFTSA.

三模块冗余(TMR)容错机制可以提供近乎完美的故障屏蔽,在提高实时系统可靠性方面具有巨大潜力。然而,一个任务的多个副本同时执行,会导致系统能耗急剧增加。在这项工作中,研究了在异构多核平台上使用 TMR 的并行应用以最小化能耗的问题。首先,改进了异构最早完成时间算法,然后根据给定应用的截止时间约束和可靠性要求,设计了一种延长副本执行时间的算法。其次,根据 TMR 的特性,设计了最小化第三副本执行开销(MEOTC)的算法。最后,考虑到任务执行的实际情况,提出了一种在线能量管理(OEM)方法。我们将所提出的算法与最先进的 AFTSA 算法进行了比较,结果表明两者在能耗方面存在显著差异。具体来说,在轻故障检测方面,MEOTC 算法和 OEM 算法的能耗分别比 AFTSA 算法低 80% 和 72%。在重故障检测方面,与 AFTSA 相比,MEOTC 和 OEM 的能耗分别为 61% 和 55%。
{"title":"Energy-efficient triple modular redundancy scheduling on heterogeneous multi-core real-time systems","authors":"Hongzhi Xu ,&nbsp;Binlian Zhang ,&nbsp;Chen Pan ,&nbsp;Keqin Li","doi":"10.1016/j.jpdc.2024.104915","DOIUrl":"https://doi.org/10.1016/j.jpdc.2024.104915","url":null,"abstract":"<div><p>Triple modular redundancy (TMR) fault tolerance mechanism can provide almost perfect fault-masking, which has the great potential to enhance the reliability of real-time systems. However, multiple copies of a task are executed concurrently, which will lead to a sharp increase in system energy consumption. In this work, the problem of parallel applications using TMR on heterogeneous multi-core platforms to minimize energy consumption is studied. First, the heterogeneous earliest finish time algorithm is improved, and then according to the given application's deadline constraints and reliability requirements, an algorithm to extend the execution time of the copies is designed. Secondly, based on the properties of TMR, an algorithm for minimizing the execution overhead of the third copy (MEOTC) is designed. Finally, considering the actual situation of task execution, an online energy management (OEM) method is proposed. The proposed algorithms were compared with the state-of-the-art AFTSA algorithm, and the results show significant differences in energy consumption. Specifically, for light fault detection, the energy consumption of the MEOTC and OEM algorithms was found to be 80% and 72% respectively, compared with AFTSA. For heavy fault detection, the energy consumption of MEOTC and OEM was measured at 61% and 55% respectively, compared with AFTSA.</p></div>","PeriodicalId":54775,"journal":{"name":"Journal of Parallel and Distributed Computing","volume":null,"pages":null},"PeriodicalIF":3.8,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140951658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Parallel and Distributed Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1