首页 > 最新文献

IEEE Transactions on Computers最新文献

英文 中文
ToEx: Accelerating Generation Stage of Transformer-Based Language Models via Token-Adaptive Early Exit ToEx:通过令牌自适应早期退出加速基于转换器的语言模型的生成阶段
IF 3.6 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-21 DOI: 10.1109/TC.2024.3404051
Myeonggu Kang;Junyoung Park;Hyein Shin;Jaekang Shin;Lee-Sup Kim
Transformer-based language models have recently gained popularity in numerous natural language processing (NLP) applications due to their superior performance compared to traditional algorithms. These models involve two execution stages: summarization and generation. The generation stage accounts for a significant portion of the total execution time due to its auto-regressive property, which necessitates considerable and repetitive off-chip accesses. Consequently, our objective is to minimize off-chip accesses during the generation stage to expedite transformer execution. To achieve the goal, we propose a token-adaptive early exit (ToEx) that generates output tokens using fewer decoders, thereby reducing off-chip accesses for loading weight parameters. Although our approach has the potential to minimize data communication, it brings two challenges: 1) inaccurate self-attention computation, and 2) significant overhead for exit decision. To overcome these challenges, we introduce a methodology that facilitates accurate self-attention by lazily performing computations for previously exited tokens. Moreover, we mitigate the overhead of exit decision by incorporating a lightweight output embedding layer. We also present a hardware design to efficiently support the proposed work. Evaluation results demonstrate that our work can reduce the number of decoders by 2.6$times$ on average. Accordingly, it achieves 3.2$times$ speedup on average compared to transformer execution without our work.
与传统算法相比,基于变换器的语言模型性能优越,因此最近在许多自然语言处理(NLP)应用中大受欢迎。这些模型涉及两个执行阶段:总结和生成。由于其自动回归特性,生成阶段占总执行时间的很大一部分,需要大量重复的片外访问。因此,我们的目标是尽量减少生成阶段的片外访问,以加快变压器的执行。为实现这一目标,我们提出了一种标记自适应早期退出(ToEx)方法,该方法使用较少的解码器生成输出标记,从而减少了用于加载权重参数的片外访问。虽然我们的方法有可能最大限度地减少数据通信,但也带来了两个挑战:1) 自我关注计算不准确,以及 2) 退出决策开销巨大。为了克服这些挑战,我们引入了一种方法,通过对先前退出的代币懒散地执行计算,促进准确的自我关注。此外,我们还通过加入轻量级输出嵌入层来减轻退出决策的开销。我们还提出了一种硬件设计,以有效支持所提出的工作。评估结果表明,我们的工作可以将解码器的数量平均减少 2.6 美元/次。因此,与没有我们的工作的变压器执行相比,它平均实现了 3.2 美元/次的提速。
{"title":"ToEx: Accelerating Generation Stage of Transformer-Based Language Models via Token-Adaptive Early Exit","authors":"Myeonggu Kang;Junyoung Park;Hyein Shin;Jaekang Shin;Lee-Sup Kim","doi":"10.1109/TC.2024.3404051","DOIUrl":"10.1109/TC.2024.3404051","url":null,"abstract":"Transformer-based language models have recently gained popularity in numerous natural language processing (NLP) applications due to their superior performance compared to traditional algorithms. These models involve two execution stages: summarization and generation. The generation stage accounts for a significant portion of the total execution time due to its auto-regressive property, which necessitates considerable and repetitive off-chip accesses. Consequently, our objective is to minimize off-chip accesses during the generation stage to expedite transformer execution. To achieve the goal, we propose a token-adaptive early exit (ToEx) that generates output tokens using fewer decoders, thereby reducing off-chip accesses for loading weight parameters. Although our approach has the potential to minimize data communication, it brings two challenges: 1) inaccurate self-attention computation, and 2) significant overhead for exit decision. To overcome these challenges, we introduce a methodology that facilitates accurate self-attention by lazily performing computations for previously exited tokens. Moreover, we mitigate the overhead of exit decision by incorporating a lightweight output embedding layer. We also present a hardware design to efficiently support the proposed work. Evaluation results demonstrate that our work can reduce the number of decoders by 2.6\u0000<inline-formula><tex-math>$times$</tex-math></inline-formula>\u0000 on average. Accordingly, it achieves 3.2\u0000<inline-formula><tex-math>$times$</tex-math></inline-formula>\u0000 speedup on average compared to transformer execution without our work.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 9","pages":"2248-2261"},"PeriodicalIF":3.6,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141147532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Relieving Write Disturbance for Phase Change Memory With RESET-Aware Data Encoding 利用 RESET 感知数据编码缓解相变存储器的写入干扰
IF 3.6 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-21 DOI: 10.1109/TC.2024.3398490
Ronglong Wu;Zhirong Shen;Jianqiang Chen;Chengshuo Zheng;Zhiwei Yang;Jiwu Shu
The write disturbance (WD) problem is becoming increasingly severe in PCM due to the continuous scaling down of memory technology. Previous studies have attempted to transform WD-vulnerable data patterns of the new data to alleviate the WD problem. However, through a wide spectrum of real-world benchmarks, we have discovered that simply transforming WD-vulnerable data patterns does not proportionally reduce (or may even increase) WD errors. To address this issue, we present ResEnc, a RESET-aware data encoding scheme that reduces RESET operations to mitigate the WD problem in both wordlines and bitlines for PCM. It dynamically establishes a mask word for each block for data encoding and adaptively selects an appropriate encoding granularity based on the diverse write patterns. ResEnc finally reassigns the mask words of unchanged blocks to changed blocks for exploring a further reduction of WD errors. Extensive experiments show that ResEnc can reduce 16.8-87.0% of WD errors, shorten 5.6-39.6% of write latency, and save 7.0-43.1% of write energy for PCM.
由于内存技术的不断升级,写入干扰(WD)问题在 PCM 中变得越来越严重。以往的研究试图转换新数据中易受 WD 影响的数据模式,以缓解 WD 问题。然而,通过广泛的实际基准测试,我们发现仅仅转换易受 WD 影响的数据模式并不能成比例地减少(甚至可能增加)WD 误差。为了解决这个问题,我们提出了一种 RESET 感知数据编码方案 ResEnc,它可以减少 RESET 操作,从而减轻 PCM 字行和比特行中的 WD 问题。它为每个数据块动态建立一个掩码字进行数据编码,并根据不同的写入模式自适应地选择适当的编码粒度。最后,ResEnc 会将未变区块的掩码字重新分配给已变区块,以进一步减少 WD 错误。大量实验表明,ResEnc 可减少 16.8%-87.0% 的 WD 错误,缩短 5.6%-39.6% 的写延迟,并为 PCM 节省 7.0%-43.1% 的写能量。
{"title":"Relieving Write Disturbance for Phase Change Memory With RESET-Aware Data Encoding","authors":"Ronglong Wu;Zhirong Shen;Jianqiang Chen;Chengshuo Zheng;Zhiwei Yang;Jiwu Shu","doi":"10.1109/TC.2024.3398490","DOIUrl":"10.1109/TC.2024.3398490","url":null,"abstract":"The write disturbance (WD) problem is becoming increasingly severe in PCM due to the continuous scaling down of memory technology. Previous studies have attempted to transform WD-vulnerable data patterns of the new data to alleviate the WD problem. However, through a wide spectrum of real-world benchmarks, we have discovered that simply transforming WD-vulnerable data patterns does not proportionally reduce (or may even increase) WD errors. To address this issue, we present \u0000<monospace>ResEnc</monospace>\u0000, a RESET-aware data encoding scheme that reduces RESET operations to mitigate the WD problem in both wordlines and bitlines for PCM. It dynamically establishes a mask word for each block for data encoding and adaptively selects an appropriate encoding granularity based on the diverse write patterns. \u0000<monospace>ResEnc</monospace>\u0000 finally reassigns the mask words of unchanged blocks to changed blocks for exploring a further reduction of WD errors. Extensive experiments show that \u0000<monospace>ResEnc</monospace>\u0000 can reduce 16.8-87.0% of WD errors, shorten 5.6-39.6% of write latency, and save 7.0-43.1% of write energy for PCM.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 8","pages":"1939-1952"},"PeriodicalIF":3.6,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141147601","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Decentralized Task Offloading in Edge Computing: An Offline-to-Online Reinforcement Learning Approach 边缘计算中的分散任务卸载:离线到在线强化学习方法
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-19 DOI: 10.1109/TC.2024.3377912
Hongcai Lin;Lei Yang;Hao Guo;Jiannong Cao
Decentralized task offloading among cooperative edge nodes has been a promising solution to enhance resource utilization and improve users’ Quality of Experience (QoE) in edge computing. However, current decentralized methods, such as heuristics and game theory-based methods, either optimize greedily or depend on rigid assumptions, failing to adapt to the dynamic edge environment. Existing DRL-based approaches train the model in a simulation and then apply it in practical systems. These methods may perform poorly because of the divergence between the practical system and the simulated environment. Other methods that train and deploy the model directly in real-world systems face a cold-start problem, which will reduce the users’ QoE before the model converges. This paper proposes a novel offline-to-online DRL called (O2O-DRL). It uses the heuristic task logs to warm-start the DRL model offline. However, offline and online data have different distributions, so using offline methods for online fine-tuning will ruin the policy learned offline. To avoid this problem, we use on-policy DRL to fine-tune the model and prevent value overestimation. We evaluate O2O-DRL with other approaches in a simulation and a Kubernetes-based testbed. The performance results show that O2O-DRL outperforms other methods and solves the cold-start problem.
在边缘计算中,合作边缘节点之间的分散式任务卸载一直是提高资源利用率和改善用户体验质量(QoE)的可行解决方案。然而,目前的分散式方法(如启发式方法和基于博弈论的方法)要么贪婪地进行优化,要么依赖于僵化的假设,无法适应动态的边缘环境。现有的基于 DRL 的方法在模拟中训练模型,然后将其应用于实际系统。由于实际系统与模拟环境之间存在差异,这些方法可能表现不佳。其他直接在实际系统中训练和部署模型的方法面临冷启动问题,这会在模型收敛之前降低用户的 QoE。本文提出了一种新颖的离线到在线 DRL(O2O-DRL)。它使用启发式任务日志来离线热启动 DRL 模型。然而,离线数据和在线数据具有不同的分布,因此使用离线方法进行在线微调会破坏离线学习的策略。为了避免这个问题,我们使用在线策略 DRL 对模型进行微调,以防止数值被高估。我们在模拟和基于 Kubernetes 的测试平台上对 O2O-DRL 和其他方法进行了评估。性能结果表明,O2O-DRL 的性能优于其他方法,并能解决冷启动问题。
{"title":"Decentralized Task Offloading in Edge Computing: An Offline-to-Online Reinforcement Learning Approach","authors":"Hongcai Lin;Lei Yang;Hao Guo;Jiannong Cao","doi":"10.1109/TC.2024.3377912","DOIUrl":"10.1109/TC.2024.3377912","url":null,"abstract":"Decentralized task offloading among cooperative edge nodes has been a promising solution to enhance resource utilization and improve users’ Quality of Experience (QoE) in edge computing. However, current decentralized methods, such as heuristics and game theory-based methods, either optimize greedily or depend on rigid assumptions, failing to adapt to the dynamic edge environment. Existing DRL-based approaches train the model in a simulation and then apply it in practical systems. These methods may perform poorly because of the divergence between the practical system and the simulated environment. Other methods that train and deploy the model directly in real-world systems face a cold-start problem, which will reduce the users’ QoE before the model converges. This paper proposes a novel offline-to-online DRL called (O2O-DRL). It uses the heuristic task logs to warm-start the DRL model offline. However, offline and online data have different distributions, so using offline methods for online fine-tuning will ruin the policy learned offline. To avoid this problem, we use on-policy DRL to fine-tune the model and prevent value overestimation. We evaluate O2O-DRL with other approaches in a simulation and a Kubernetes-based testbed. The performance results show that O2O-DRL outperforms other methods and solves the cold-start problem.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 6","pages":"1603-1615"},"PeriodicalIF":3.7,"publicationDate":"2024-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140170080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HPDK: A Hybrid PM-DRAM Key-Value Store for High I/O Throughput HPDK:实现高 I/O 吞吐量的混合 PM-DRAM 键值存储器
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-18 DOI: 10.1109/TC.2024.3377914
Bihui Liu;Zhenyu Ye;Qiao Hu;Yupeng Hu;Yuchong Hu;Yang Xu;Keqin Li
This paper explores the design of an architecture that replaces Disk with Persistent Memory (PM) to achieve the highest I/O throughput in Log-Structured Merge Tree (LSM-Tree) based key-value stores (KVS). Most existing LSM-Tree based KVSs use PM as an intermediate or smoothing layer, which fails to fully exploit PM's unique advantages to maximize I/O throughput. However, due to PM's distinct characteristics, such as byte addressability and short erasure time, simply replacing existing storage with PM does not yield optimal I/O performance. Furthermore, LSM-Tree based KVSs often face slow read performance. To tackle these challenges, this paper presents HPDK, a hybrid PM-DRAM KVS that combines level compression for LSM-Trees in PM with a B${}^{+}$-tree based in-memory search index in DRAM, resulting in high write and read throughput. HPDK also employs a key-value separation design and a live-item rate-based dynamic merge method to reduce the volume of PM writes. We implement and evaluate HPDK using a real PM drive, and our extensive experiments show that HPDK provides 1.25-11.8 and 1.47-36.4 times higher read and write throughput, respectively, compared to other state-of-the-art LSM-Tree based approaches.
本文探讨了用持久内存(PM)取代磁盘的架构设计,以在基于日志结构合并树(LSM-Tree)的键值存储(KVS)中实现最高的 I/O 吞吐量。现有的大多数基于 LSM-Tree 的 KVS 都将 PM 用作中间层或平滑层,未能充分利用 PM 的独特优势来最大限度地提高 I/O 吞吐量。然而,由于 PM 具有字节寻址能力和短擦除时间等显著特点,简单地用 PM 取代现有存储并不能获得最佳 I/O 性能。此外,基于 LSM-Tree 的 KVS 经常面临读取性能缓慢的问题。为了应对这些挑战,本文提出了 HPDK,一种混合 PM-DRAM KVS,它将 PM 中 LSM 树的级别压缩与 DRAM 中基于 B${}^{+}$ 树的内存搜索索引相结合,从而实现了高写入和读取吞吐量。HPDK 还采用了键值分离设计和基于实时项速率的动态合并方法,以减少 PM 的写入量。我们使用实际的 PM 驱动器实现并评估了 HPDK,大量实验表明,与其他基于 LSM-Tree 的先进方法相比,HPDK 的读取和写入吞吐量分别高出 1.25-11.8 倍和 1.47-36.4 倍。
{"title":"HPDK: A Hybrid PM-DRAM Key-Value Store for High I/O Throughput","authors":"Bihui Liu;Zhenyu Ye;Qiao Hu;Yupeng Hu;Yuchong Hu;Yang Xu;Keqin Li","doi":"10.1109/TC.2024.3377914","DOIUrl":"10.1109/TC.2024.3377914","url":null,"abstract":"This paper explores the design of an architecture that replaces Disk with Persistent Memory (PM) to achieve the highest I/O throughput in Log-Structured Merge Tree (LSM-Tree) based key-value stores (KVS). Most existing LSM-Tree based KVSs use PM as an intermediate or smoothing layer, which fails to fully exploit PM's unique advantages to maximize I/O throughput. However, due to PM's distinct characteristics, such as byte addressability and short erasure time, simply replacing existing storage with PM does not yield optimal I/O performance. Furthermore, LSM-Tree based KVSs often face slow read performance. To tackle these challenges, this paper presents HPDK, a hybrid PM-DRAM KVS that combines level compression for LSM-Trees in PM with a B\u0000<inline-formula><tex-math>${}^{+}$</tex-math></inline-formula>\u0000-tree based in-memory search index in DRAM, resulting in high write and read throughput. HPDK also employs a key-value separation design and a live-item rate-based dynamic merge method to reduce the volume of PM writes. We implement and evaluate HPDK using a real PM drive, and our extensive experiments show that HPDK provides 1.25-11.8 and 1.47-36.4 times higher read and write throughput, respectively, compared to other state-of-the-art LSM-Tree based approaches.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 6","pages":"1575-1587"},"PeriodicalIF":3.7,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140170171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Big-PERCIVAL: Exploring the Native Use of 64-Bit Posit Arithmetic in Scientific Computing Big-PERCIVAL:探索科学计算中 64 位正则表达式的本地使用
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-18 DOI: 10.1109/TC.2024.3377890
David Mallasén;Alberto A. Del Barrio;Manuel Prieto-Matias
The accuracy requirements in many scientific computing workloads result in the use of double-precision floating-point arithmetic in the execution kernels. Nevertheless, emerging real-number representations, such as posit arithmetic, show promise in delivering even higher accuracy in such computations. In this work, we explore the native use of 64-bit posits in a series of numerical benchmarks and compare their timing performance, accuracy and hardware cost to IEEE 754 doubles. In addition, we also study the conjugate gradient method for numerically solving systems of linear equations in real-world applications. For this, we extend the PERCIVAL RISC-V core and the Xposit custom RISC-V extension with posit64 and quire operations. Results show that posit64 can obtain up to 4 orders of magnitude lower mean square error than doubles. This leads to a reduction in the number of iterations required for convergence in some iterative solvers. However, leveraging the quire accumulator register can limit the order of some operations such as matrix multiplications. Furthermore, detailed FPGA and ASIC synthesis results highlight the significant hardware cost of 64-bit posit arithmetic and quire. Despite this, the large accuracy improvements achieved with the same memory bandwidth suggest that posit arithmetic may provide a potential alternative representation for scientific computing.
许多科学计算工作负载对精度的要求导致在执行内核中使用双精度浮点运算。不过,新出现的实数表示法(如正位运算)有望在此类计算中提供更高的精度。在这项工作中,我们在一系列数值基准中探索了 64 位实数的本地使用,并将其时序性能、精度和硬件成本与 IEEE 754 双倍进行了比较。此外,我们还研究了共轭梯度法,用于数值求解实际应用中的线性方程组。为此,我们使用 posit64 和 quire 运算扩展了 PERCIVAL RISC-V 内核和 Xposit 定制 RISC-V 扩展。结果表明,posit64 的均方误差比 double 低 4 个数量级。这减少了某些迭代求解器收敛所需的迭代次数。不过,利用 quire 累加器寄存器可能会限制矩阵乘法等某些运算的阶次。此外,详细的 FPGA 和 ASIC 综合结果表明,64 位正则运算和 quire 的硬件成本很高。尽管如此,在相同内存带宽下实现的巨大精度提升表明,posit 运算可能为科学计算提供一种潜在的替代表示方法。
{"title":"Big-PERCIVAL: Exploring the Native Use of 64-Bit Posit Arithmetic in Scientific Computing","authors":"David Mallasén;Alberto A. Del Barrio;Manuel Prieto-Matias","doi":"10.1109/TC.2024.3377890","DOIUrl":"10.1109/TC.2024.3377890","url":null,"abstract":"The accuracy requirements in many scientific computing workloads result in the use of double-precision floating-point arithmetic in the execution kernels. Nevertheless, emerging real-number representations, such as posit arithmetic, show promise in delivering even higher accuracy in such computations. In this work, we explore the native use of 64-bit posits in a series of numerical benchmarks and compare their timing performance, accuracy and hardware cost to IEEE 754 doubles. In addition, we also study the conjugate gradient method for numerically solving systems of linear equations in real-world applications. For this, we extend the PERCIVAL RISC-V core and the Xposit custom RISC-V extension with posit64 and quire operations. Results show that posit64 can obtain up to 4 orders of magnitude lower mean square error than doubles. This leads to a reduction in the number of iterations required for convergence in some iterative solvers. However, leveraging the quire accumulator register can limit the order of some operations such as matrix multiplications. Furthermore, detailed FPGA and ASIC synthesis results highlight the significant hardware cost of 64-bit posit arithmetic and quire. Despite this, the large accuracy improvements achieved with the same memory bandwidth suggest that posit arithmetic may provide a potential alternative representation for scientific computing.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 6","pages":"1472-1485"},"PeriodicalIF":3.7,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10473215","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140169609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Dynamic Adaptive Framework for Practical Byzantine Fault Tolerance Consensus Protocol in the Internet of Things 物联网中实用拜占庭容错共识协议的动态自适应框架
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-18 DOI: 10.1109/TC.2024.3377921
Chunpei Li;Wangjie Qiu;Xianxian Li;Chen Liu;Zhiming Zheng
The Practical Byzantine Fault Tolerance (PBFT) protocol-supported blockchain can provide decentralized security and trust mechanisms for the Internet of Things (IoT). However, the PBFT protocol is not specifically designed for IoT applications. Consequently, adapting PBFT to the dynamic changes of an IoT environment with incomplete information represents a challenge that urgently needs to be addressed. To this end, we introduce DA-PBFT, a PBFT dynamic adaptive framework based on a multi-agent architecture. DA-PBFT divides the dynamic adaptive process into two sub-processes: optimality-seeking and optimization decision-making. During the optimality-seeking process, a PBFT optimization model is constructed based on deep reinforcement learning. This model is designed to generate PBFT optimization strategies for consensus nodes. In the optimization decision-making process, a PBFT optimization decision consensus mechanism is constructed based on the Borda count method. This mechanism ensures consistency in PBFT optimization decisions within an environment characterized by incomplete information. Furthermore, we designed a dynamic adaptive incentive mechanism to explore the Nash equilibrium conditions and security aspects of DA-PBFT. The experimental results demonstrate that DA-PBFT is capable of achieving consistency in PBFT optimization decisions within an environment of incomplete information, thereby offering robust and efficient transaction throughput for IoT applications.
由实用拜占庭容错(PBFT)协议支持的区块链可以为物联网(IoT)提供去中心化的安全和信任机制。然而,PBFT 协议并非专为物联网应用而设计。因此,如何使 PBFT 适应信息不完整的物联网环境的动态变化是一个亟待解决的挑战。为此,我们引入了基于多代理架构的 PBFT 动态自适应框架 DA-PBFT。DA-PBFT 将动态自适应过程分为两个子过程:优化搜索和优化决策。在寻求优化过程中,基于深度强化学习构建了一个 PBFT 优化模型。该模型旨在为共识节点生成 PBFT 优化策略。在优化决策过程中,基于博尔达计数法构建了 PBFT 优化决策共识机制。该机制确保了 PBFT 优化决策在不完全信息环境中的一致性。此外,我们还设计了一种动态自适应激励机制,以探索 DA-PBFT 的纳什均衡条件和安全性。实验结果表明,DA-PBFT 能够在不完全信息环境中实现 PBFT 优化决策的一致性,从而为物联网应用提供稳健高效的交易吞吐量。
{"title":"A Dynamic Adaptive Framework for Practical Byzantine Fault Tolerance Consensus Protocol in the Internet of Things","authors":"Chunpei Li;Wangjie Qiu;Xianxian Li;Chen Liu;Zhiming Zheng","doi":"10.1109/TC.2024.3377921","DOIUrl":"10.1109/TC.2024.3377921","url":null,"abstract":"The Practical Byzantine Fault Tolerance (PBFT) protocol-supported blockchain can provide decentralized security and trust mechanisms for the Internet of Things (IoT). However, the PBFT protocol is not specifically designed for IoT applications. Consequently, adapting PBFT to the dynamic changes of an IoT environment with incomplete information represents a challenge that urgently needs to be addressed. To this end, we introduce DA-PBFT, a PBFT dynamic adaptive framework based on a multi-agent architecture. DA-PBFT divides the dynamic adaptive process into two sub-processes: optimality-seeking and optimization decision-making. During the optimality-seeking process, a PBFT optimization model is constructed based on deep reinforcement learning. This model is designed to generate PBFT optimization strategies for consensus nodes. In the optimization decision-making process, a PBFT optimization decision consensus mechanism is constructed based on the Borda count method. This mechanism ensures consistency in PBFT optimization decisions within an environment characterized by incomplete information. Furthermore, we designed a dynamic adaptive incentive mechanism to explore the Nash equilibrium conditions and security aspects of DA-PBFT. The experimental results demonstrate that DA-PBFT is capable of achieving consistency in PBFT optimization decisions within an environment of incomplete information, thereby offering robust and efficient transaction throughput for IoT applications.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 7","pages":"1669-1682"},"PeriodicalIF":3.7,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140169591","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Integrated FPGA Accelerator for Deep Learning-Based 2D/3D Path Planning 基于深度学习的 2D/3D 路径规划的 FPGA 集成加速器
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-18 DOI: 10.1109/TC.2024.3377895
Keisuke Sugiura;Hiroki Matsutani
Path planning is a crucial component for realizing the autonomy of mobile robots. However, due to limited computational resources on mobile robots, it remains challenging to deploy state-of-the-art methods and achieve real-time performance. To address this, we propose P3Net (PointNet-based Path Planning Networks), a lightweight deep-learning-based method for 2D/3D path planning, and design an IP core (P3NetCore) targeting FPGA SoCs (Xilinx ZCU104). P3Net improves the algorithm and model architecture of the recently-proposed MPNet. P3Net employs an encoder with a PointNet backbone and a lightweight planning network in order to extract robust point cloud features and sample path points from a promising region. P3NetCore is comprised of the fully-pipelined point cloud encoder, batched bidirectional path planner, and parallel collision checker, to cover most part of the algorithm. On the 2D (3D) datasets, P3Net with the IP core runs 30.52–186.36x and 7.68–143.62x (15.69–93.26x and 5.30–45.27x) faster than ARM Cortex CPU and Nvidia Jetson while only consuming 0.255W (0.809W), and is up to 1278.14x (455.34x) power-efficient than the workstation. P3Net improves the success rate by up to 28.2% and plans a near-optimal path, leading to a significantly better tradeoff between computation and solution quality than MPNet and the state-of-the-art sampling-based methods.
路径规划是实现移动机器人自主性的关键组成部分。然而,由于移动机器人的计算资源有限,部署最先进的方法并实现实时性能仍具有挑战性。为解决这一问题,我们提出了基于深度学习的轻量级 2D/3D 路径规划方法 P3Net(基于 PointNet 的路径规划网络),并设计了针对 FPGA SoC(赛灵思 ZCU104)的 IP 核(P3NetCore)。P3Net 改进了最近提出的 MPNet 的算法和模型架构。P3Net 采用带有 PointNet 主干网和轻量级规划网的编码器,以便从有希望的区域提取稳健的点云特征和路径点样本。P3NetCore 由全管道点云编码器、批量双向路径规划器和并行碰撞检查器组成,涵盖了算法的大部分内容。在二维(三维)数据集上,IP 核的 P3Net 运行速度是 ARM Cortex CPU 和 Nvidia Jetson 的 30.52-186.36 倍和 7.68-143.62 倍(15.69-93.26 倍和 5.30-45.27 倍),而功耗仅为 0.255W (0.809W),是工作站的 1278.14 倍(455.34 倍)。P3Net 将成功率提高了 28.2%,并规划了一条接近最优的路径,从而在计算和解决方案质量之间实现了明显优于 MPNet 和最先进的基于采样的方法的权衡。
{"title":"An Integrated FPGA Accelerator for Deep Learning-Based 2D/3D Path Planning","authors":"Keisuke Sugiura;Hiroki Matsutani","doi":"10.1109/TC.2024.3377895","DOIUrl":"10.1109/TC.2024.3377895","url":null,"abstract":"Path planning is a crucial component for realizing the autonomy of mobile robots. However, due to limited computational resources on mobile robots, it remains challenging to deploy state-of-the-art methods and achieve real-time performance. To address this, we propose P3Net (PointNet-based Path Planning Networks), a lightweight deep-learning-based method for 2D/3D path planning, and design an IP core (P3NetCore) targeting FPGA SoCs (Xilinx ZCU104). P3Net improves the algorithm and model architecture of the recently-proposed MPNet. P3Net employs an encoder with a PointNet backbone and a lightweight planning network in order to extract robust point cloud features and sample path points from a promising region. P3NetCore is comprised of the fully-pipelined point cloud encoder, batched bidirectional path planner, and parallel collision checker, to cover most part of the algorithm. On the 2D (3D) datasets, P3Net with the IP core runs 30.52–186.36x and 7.68–143.62x (15.69–93.26x and 5.30–45.27x) faster than ARM Cortex CPU and Nvidia Jetson while only consuming 0.255W (0.809W), and is up to 1278.14x (455.34x) power-efficient than the workstation. P3Net improves the success rate by up to 28.2% and plans a near-optimal path, leading to a significantly better tradeoff between computation and solution quality than MPNet and the state-of-the-art sampling-based methods.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 6","pages":"1442-1456"},"PeriodicalIF":3.7,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10474486","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140169822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reordering and Compression for Hypergraph Processing 超图处理的重新排序与压缩
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-18 DOI: 10.1109/TC.2024.3377915
Yu Liu;Qi Luo;Mengbai Xiao;Dongxiao Yu;Huashan Chen;Xiuzhen Cheng
Hypergraphs are applicable to various domains such as social contagion, online groups, and protein structures due to their effective modeling of multivariate relationships. However, the increasing size of hypergraphs has led to high computation costs, necessitating efficient acceleration strategies. Existing approaches often require consideration of algorithm-specific issues, making them difficult to directly apply to arbitrary hypergraph processing tasks. In this paper, we propose a compression-array acceleration strategy involving hypergraph reordering to improve memory access efficiency, which can be applied to various hypergraph processing tasks without considering the algorithm itself. We introduce a new metric called closeness to optimize the ordering of vertices and hyperedges in the one-dimensional array representation. Moreover, we present an $frac{1}{2w}$-approximation algorithm to obtain the optimal ordering of vertices and hyperedges. We also develop an efficient update mechanism for dynamic hypergraphs. Our extensive experiments demonstrate significant improvements in hypergraph processing performance, reduced cache misses, and reduced memory footprint. Furthermore, our method can be integrated into existing hypergraph processing frameworks, such as Hygra, to enhance their performance.
超图可有效模拟多元关系,因此适用于社会传染、在线群组和蛋白质结构等多个领域。然而,超图的规模越来越大,导致计算成本越来越高,因此需要高效的加速策略。现有方法通常需要考虑特定算法的问题,因此难以直接应用于任意超图处理任务。在本文中,我们提出了一种涉及超图重新排序的压缩阵列加速策略,以提高内存访问效率,该策略可应用于各种超图处理任务,而无需考虑算法本身。我们引入了一种名为 "接近度 "的新度量来优化一维数组表示中顶点和超图的排序。此外,我们还提出了一种 $frac{1}{2w}$ 近似算法来获得顶点和超边的最优排序。我们还为动态超图开发了一种高效的更新机制。我们的大量实验证明,超图处理性能有了显著提高,高速缓存缺失减少,内存占用降低。此外,我们的方法可以集成到现有的超图处理框架(如 Hygra)中,以提高其性能。
{"title":"Reordering and Compression for Hypergraph Processing","authors":"Yu Liu;Qi Luo;Mengbai Xiao;Dongxiao Yu;Huashan Chen;Xiuzhen Cheng","doi":"10.1109/TC.2024.3377915","DOIUrl":"10.1109/TC.2024.3377915","url":null,"abstract":"Hypergraphs are applicable to various domains such as social contagion, online groups, and protein structures due to their effective modeling of multivariate relationships. However, the increasing size of hypergraphs has led to high computation costs, necessitating efficient acceleration strategies. Existing approaches often require consideration of algorithm-specific issues, making them difficult to directly apply to arbitrary hypergraph processing tasks. In this paper, we propose a compression-array acceleration strategy involving hypergraph reordering to improve memory access efficiency, which can be applied to various hypergraph processing tasks without considering the algorithm itself. We introduce a new metric called closeness to optimize the ordering of vertices and hyperedges in the one-dimensional array representation. Moreover, we present an \u0000<inline-formula><tex-math>$frac{1}{2w}$</tex-math></inline-formula>\u0000-approximation algorithm to obtain the optimal ordering of vertices and hyperedges. We also develop an efficient update mechanism for dynamic hypergraphs. Our extensive experiments demonstrate significant improvements in hypergraph processing performance, reduced cache misses, and reduced memory footprint. Furthermore, our method can be integrated into existing hypergraph processing frameworks, such as Hygra, to enhance their performance.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 6","pages":"1486-1499"},"PeriodicalIF":3.7,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140169603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prefender: A Prefetching Defender Against Cache Side Channel Attacks as a Pretender Prefender:作为伪装者抵御缓存侧通道攻击的预取防御器
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-18 DOI: 10.1109/TC.2024.3377891
Luyi Li;Jiayi Huang;Lang Feng;Zhongfeng Wang
Cache side channel attacks are increasingly alarming in modern processors due to the recent emergence of Spectre and Meltdown attacks. A typical attack performs intentional cache access and manipulates cache states to leak secrets by observing the victim's cache access patterns. Different countermeasures have been proposed to defend against both general and transient execution based attacks. Despite their effectiveness, they mostly trade some level of performance for security, or have restricted security scope. In this paper, we seek an approach to enforcing security while maintaining performance. We leverage the insight that attackers need to access cache in order to manipulate and observe cache state changes for information leakage. Specifically, we propose Prefender, a secure prefetcher that learns and predicts attack-related accesses for prefetching the cachelines to simultaneously help security and performance. Our results show that Prefender is effective against several cache side channel attacks while maintaining or even improving performance for SPEC CPU 2006 and 2017 benchmarks.
由于最近出现的 Spectre 和 Meltdown 攻击,现代处理器中的高速缓存侧信道攻击越来越令人担忧。典型的攻击是通过观察受害者的高速缓存访问模式,执行有意的高速缓存访问并操纵高速缓存状态以泄露机密。针对基于一般执行和瞬态执行的攻击,人们提出了不同的应对措施。尽管这些对策很有效,但它们大多以一定程度的性能来换取安全性,或者安全范围有限。在本文中,我们寻求一种既能保证安全性,又能保持性能的方法。我们利用攻击者需要访问高速缓存才能操纵和观察高速缓存状态变化以防止信息泄漏这一观点。具体来说,我们提出了一种安全预取器--Prefender,它可以学习和预测与攻击相关的访问,以便预取缓存行,从而同时提高安全性和性能。我们的研究结果表明,在 SPEC CPU 2006 和 2017 基准测试中,Prefender 能有效抵御多种缓存侧信道攻击,同时保持甚至提高性能。
{"title":"Prefender: A Prefetching Defender Against Cache Side Channel Attacks as a Pretender","authors":"Luyi Li;Jiayi Huang;Lang Feng;Zhongfeng Wang","doi":"10.1109/TC.2024.3377891","DOIUrl":"10.1109/TC.2024.3377891","url":null,"abstract":"Cache side channel attacks are increasingly alarming in modern processors due to the recent emergence of Spectre and Meltdown attacks. A typical attack performs intentional cache access and manipulates cache states to leak secrets by observing the victim's cache access patterns. Different countermeasures have been proposed to defend against both general and transient execution based attacks. Despite their effectiveness, they mostly trade some level of performance for security, or have restricted security scope. In this paper, we seek an approach to enforcing security while maintaining performance. We leverage the insight that attackers need to access cache in order to manipulate and observe cache state changes for information leakage. Specifically, we propose \u0000<sc>Prefender</small>\u0000, a secure prefetcher that learns and predicts attack-related accesses for prefetching the cachelines to simultaneously help security and performance. Our results show that \u0000<sc>Prefender</small>\u0000 is effective against several cache side channel attacks while maintaining or even improving performance for SPEC CPU 2006 and 2017 benchmarks.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 6","pages":"1457-1471"},"PeriodicalIF":3.7,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140169588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ElasticDNN: On-Device Neural Network Remodeling for Adapting Evolving Vision Domains at Edge ElasticDNN:在设备上重塑神经网络,以适应边缘不断变化的视觉领域
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-14 DOI: 10.1109/TC.2024.3375608
Qinglong Zhang;Rui Han;Chi Harold Liu;Guoren Wang;Lydia Y. Chen
Executing deep neural networks (DNN) based vision tasks on edge devices encounters challenging scenarios of significant and continually evolving data domains (e.g. background or subpopulation shift). With limited resources, the state-of-the-art domain adaptation (DA) methods either cause high training overheads on large DNN models, or incur significant accuracy losses when adapting small/compressed models in an online fashion. The inefficient resource scheduling among multiple applications further degrades their overall model accuracy. In this paper, we present ElasticDNN, a framework that enables online DNN remodeling for applications encountering evolving domain drifts at edge. Its first key component is the master-surrogate DNN models, which can dynamically generate a small surrogate DNN by retaining and training the large master DNN's most relevant regions pertinent to the new domain. The second novelty of ElasticDNN is the filter-grained resource scheduling, which allocates GPU resources based on online accuracy estimation and DNN remodeling of co-running applications. We fully implement ElasticDNN and demonstrate its effectiveness through extensive experiments. The results show that, compared to existing online DA methods using the same model sizes, ElasticDNN improves accuracy by 23.31% and reduces adaption time by 35.67x. In the more challenging multi-application scenario, ElasticDNN improves accuracy by an average of 25.91%.
在边缘设备上执行基于深度神经网络(DNN)的视觉任务时,会遇到数据域(如背景或子群体变化)显著且不断变化的挑战性场景。在资源有限的情况下,最先进的域自适应(DA)方法要么会对大型 DNN 模型造成较高的训练开销,要么会在以在线方式自适应小型/压缩模型时造成显著的精度损失。多个应用之间低效的资源调度进一步降低了模型的整体准确性。在本文中,我们介绍了 ElasticDNN,这是一个能为遇到不断变化的边缘领域漂移的应用实现在线 DNN 重塑的框架。它的第一个关键组件是主代理 DNN 模型,它可以通过保留和训练大型主 DNN 中与新领域相关的最相关区域,动态生成小型代理 DNN。ElasticDNN 的第二个创新点是过滤粒度资源调度,它根据在线准确性评估和共同运行应用程序的 DNN 重塑情况来分配 GPU 资源。我们完全实现了 ElasticDNN,并通过大量实验证明了其有效性。结果表明,与使用相同模型大小的现有在线 DA 方法相比,ElasticDNN 将准确率提高了 23.31%,并将适应时间缩短了 35.67 倍。在更具挑战性的多应用场景中,ElasticDNN 平均提高了 25.91% 的准确率。
{"title":"ElasticDNN: On-Device Neural Network Remodeling for Adapting Evolving Vision Domains at Edge","authors":"Qinglong Zhang;Rui Han;Chi Harold Liu;Guoren Wang;Lydia Y. Chen","doi":"10.1109/TC.2024.3375608","DOIUrl":"10.1109/TC.2024.3375608","url":null,"abstract":"Executing deep neural networks (DNN) based vision tasks on edge devices encounters challenging scenarios of significant and continually evolving data domains (e.g. background or subpopulation shift). With limited resources, the state-of-the-art domain adaptation (DA) methods either cause high training overheads on large DNN models, or incur significant accuracy losses when adapting small/compressed models in an online fashion. The inefficient resource scheduling among multiple applications further degrades their overall model accuracy. In this paper, we present ElasticDNN, a framework that enables online DNN remodeling for applications encountering evolving domain drifts at edge. Its first key component is the master-surrogate DNN models, which can dynamically generate a small surrogate DNN by retaining and training the large master DNN's most relevant regions pertinent to the new domain. The second novelty of ElasticDNN is the filter-grained resource scheduling, which allocates GPU resources based on online accuracy estimation and DNN remodeling of co-running applications. We fully implement ElasticDNN and demonstrate its effectiveness through extensive experiments. The results show that, compared to existing online DA methods using the same model sizes, ElasticDNN improves accuracy by 23.31% and reduces adaption time by 35.67x. In the more challenging multi-application scenario, ElasticDNN improves accuracy by an average of 25.91%.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 6","pages":"1616-1630"},"PeriodicalIF":3.7,"publicationDate":"2024-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140155045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Computers
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1