首页 > 最新文献

IEEE Transactions on Computers最新文献

英文 中文
Un-IOV: Achieving Bare-Metal Level I/O Virtualization Performance for Cloud Usage With Migratability, Scalability and Transparency Un-IOV:通过可迁移性、可扩展性和透明度,为云使用实现裸机级 I/O 虚拟化性能
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-14 DOI: 10.1109/TC.2024.3375589
Zongpu Zhang;Chenbo Xia;Cunming Liang;Jian Li;Chen Yu;Tiwei Bie;Roberts Martin;Daly Dan;Xiao Wang;Yong Liu;Haibing Guan
I/O virtualization is utilized by cloud platforms to provide tenants with efficient, scalable, and manageable network and storage services. The de-facto industrial standard, paravirtualization, offers rich cloud functionality by introducing split front-end and back-end drivers in the guest and host operating systems, respectively. Given this fact, paravirtualization incurs host inefficiency and performance overhead. Thus, emerging hardware virtio accelerators (i.e., SRIOV-capable devices that conform to virtio specification) with device passthrough technologies mitigate the performance issue. However, adopting these devices presents the challenge of insufficient support for live migration. This paper proposes Un-IOV, a novel I/O virtualization system that simultaneously achieves bare-metal level I/O performance and migratability. The key idea is to develop a new hybrid virtualization stack with: (1) a host-bypassed direct data path for virtio accelerators, and (2) a relayed control path guaranteeing seamless live migration support. Un-IOV achieves high scalability by consuming minimum host resources. Extensive experiment results demonstrate that Un-IOV achieves superior network and storage virtualization performance than software implementations with comparable performance of direct passthrough I/O virtualization, while imposing zero guest modification (i.e., guest transparency).
云平台利用 I/O 虚拟化为租户提供高效、可扩展、可管理的网络和存储服务。事实上的行业标准--准虚拟化,通过在客户和主机操作系统中分别引入分离的前端和后端驱动程序,提供了丰富的云功能。鉴于此,准虚拟化会造成主机效率低下和性能开销。因此,采用设备直通技术的新兴硬件 virtio 加速器(即符合 virtio 规范的 SRIOV 兼容设备)可以缓解性能问题。然而,采用这些设备会面临对实时迁移支持不足的挑战。本文提出的 Un-IOV 是一种新型 I/O 虚拟化系统,可同时实现裸机级 I/O 性能和可迁移性。其主要思路是开发一种新的混合虚拟化堆栈,其中包括(1) 用于 virtio 加速器的主机旁路直接数据路径,以及 (2) 保证无缝实时迁移支持的中继控制路径。Un-IOV 通过消耗最少的主机资源实现了高可扩展性。广泛的实验结果表明,Un-IOV 实现的网络和存储虚拟化性能优于直接直通 I/O 虚拟化性能相当的软件实现,同时对客户机零修改(即客户机透明)。
{"title":"Un-IOV: Achieving Bare-Metal Level I/O Virtualization Performance for Cloud Usage With Migratability, Scalability and Transparency","authors":"Zongpu Zhang;Chenbo Xia;Cunming Liang;Jian Li;Chen Yu;Tiwei Bie;Roberts Martin;Daly Dan;Xiao Wang;Yong Liu;Haibing Guan","doi":"10.1109/TC.2024.3375589","DOIUrl":"10.1109/TC.2024.3375589","url":null,"abstract":"I/O virtualization is utilized by cloud platforms to provide tenants with efficient, scalable, and manageable network and storage services. The de-facto industrial standard, paravirtualization, offers rich cloud functionality by introducing split front-end and back-end drivers in the guest and host operating systems, respectively. Given this fact, paravirtualization incurs host inefficiency and performance overhead. Thus, emerging hardware virtio accelerators (i.e., SRIOV-capable devices that conform to virtio specification) with device passthrough technologies mitigate the performance issue. However, adopting these devices presents the challenge of insufficient support for live migration. This paper proposes Un-IOV, a novel I/O virtualization system that simultaneously achieves bare-metal level I/O performance and migratability. The key idea is to develop a new hybrid virtualization stack with: (1) a host-bypassed direct data path for virtio accelerators, and (2) a relayed control path guaranteeing seamless live migration support. Un-IOV achieves high scalability by consuming minimum host resources. Extensive experiment results demonstrate that Un-IOV achieves superior network and storage virtualization performance than software implementations with comparable performance of direct passthrough I/O virtualization, while imposing zero guest modification (i.e., guest transparency).","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 7","pages":"1655-1668"},"PeriodicalIF":3.7,"publicationDate":"2024-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140154872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FPGA-Accelerated Range-Limited Molecular Dynamics FPGA 加速的限幅分子动力学
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-14 DOI: 10.1109/TC.2024.3375613
Chunshu Wu;Chen Yang;Sahan Bandara;Tong Geng;Anqi Guo;Pouya Haghi;Ang Li;Martin Herbordt
Long timescale Molecular Dynamics (MD) simulation of small molecules is crucial in drug design and basic science. To accelerate a small data set that is executed for a large number of iterations, high-efficiency is required. Recent work in this domain has demonstrated that among COTS devices only FPGA-centric clusters can scale beyond a few processors. The problem addressed here is that, as the number of on-chip processors has increased from fewer than 10 into the hundreds, previous intra-chip routing solutions are no longer viable. We find, however, that through various design innovations, high efficiency can be maintained. These include replacing the previous broadcast networks with ring-routing and then augmenting the rings with out-of-order and caching mechanisms. Others are adding a level of hierarchical filtering and memory recycling. Two novel optimized architectures emerge, together with a number of variations. These are validated, analyzed, and evaluated. We find that in the domain of interest speed-ups over GPUs are achieved. The potential impact is that this system promises to be the basis for scalable long timescale MD with commodity clusters.
小分子的长时间尺度分子动力学(MD)模拟对药物设计和基础科学至关重要。要加速执行大量迭代的小数据集,就需要高效率。该领域的最新研究表明,在 COTS 设备中,只有以 FPGA 为中心的集群可以扩展到几个处理器以上。这里要解决的问题是,随着片上处理器数量从不到 10 个增加到数百个,以前的片内路由解决方案已不再可行。但我们发现,通过各种设计创新,可以保持高效率。这些创新包括用环形路由取代以前的广播网络,然后用失序和缓存机制增强环形路由。其他创新还包括增加分层过滤和内存循环。在此基础上,出现了两种新颖的优化架构以及多种变体。我们对这些架构进行了验证、分析和评估。我们发现,在我们感兴趣的领域中,速度比 GPU 更快。其潜在影响是,该系统有望成为使用商品集群进行可扩展的长时间尺度 MD 的基础。
{"title":"FPGA-Accelerated Range-Limited Molecular Dynamics","authors":"Chunshu Wu;Chen Yang;Sahan Bandara;Tong Geng;Anqi Guo;Pouya Haghi;Ang Li;Martin Herbordt","doi":"10.1109/TC.2024.3375613","DOIUrl":"10.1109/TC.2024.3375613","url":null,"abstract":"Long timescale Molecular Dynamics (MD) simulation of small molecules is crucial in drug design and basic science. To accelerate a small data set that is executed for a large number of iterations, high-efficiency is required. Recent work in this domain has demonstrated that among COTS devices only FPGA-centric clusters can scale beyond a few processors. The problem addressed here is that, as the number of on-chip processors has increased from fewer than 10 into the hundreds, previous intra-chip routing solutions are no longer viable. We find, however, that through various design innovations, high efficiency can be maintained. These include replacing the previous broadcast networks with ring-routing and then augmenting the rings with out-of-order and caching mechanisms. Others are adding a level of hierarchical filtering and memory recycling. Two novel optimized architectures emerge, together with a number of variations. These are validated, analyzed, and evaluated. We find that in the domain of interest speed-ups over GPUs are achieved. The potential impact is that this system promises to be the basis for scalable long timescale MD with commodity clusters.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 6","pages":"1544-1558"},"PeriodicalIF":3.7,"publicationDate":"2024-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140154879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AdaptMD: Balancing Space and Performance in NUMA Architectures With Adaptive Memory Deduplication AdaptMD:利用自适应内存重复数据删除技术平衡 NUMA 架构的空间和性能
IF 3.7 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-14 DOI: 10.1109/TC.2024.3375592
Lulu Yao;Yongkun Li;Patrick P. C. Lee;Xiaoyang Wang;Yinlong Xu
Memory deduplication effectively relieves the memory space bottleneck by removing duplicate pages, especially in virtualized systems in which virtual machines run the same OS and similar applications. However, due to the non-uniform access latencies in NUMA architectures, memory deduplication poses a trade-off between memory savings and access performance: global deduplication across NUMA nodes realizes high memory savings, but leads to frequent cross-node remote access after deduplication and results in performance degradations. In contrast, local deduplication avoids remote access, but limits deduplication effectiveness. We design AdaptMD, an adaptive memory deduplication system that addresses the space-performance trade-off in NUMA architectures. AdaptMD leverages hotness awareness to globally deduplicate only cold pages to reduce remote access. It also migrates similar applications to the same NUMA node to allow local deduplication without remote access. We further make AdaptMD readily configurable to address various deployment scenarios. Experiments show that AdaptMD achieves high memory savings as in global deduplication, while achieving similar access performance as in local deduplication.
重复内存删除通过删除重复页面有效缓解了内存空间瓶颈,尤其是在虚拟机运行相同操作系统和类似应用程序的虚拟化系统中。然而,由于 NUMA 架构的访问延迟不均匀,重复数据删除需要在节省内存和访问性能之间做出权衡:跨 NUMA 节点的全局重复数据删除可节省大量内存,但会导致重复数据删除后频繁的跨节点远程访问,从而导致性能下降。相比之下,本地重复数据删除避免了远程访问,但却限制了重复数据删除的效果。我们设计的 AdaptMD 是一种自适应重复数据删除内存系统,可解决 NUMA 架构中空间与性能之间的权衡问题。AdaptMD 利用热度感知功能,只对冷页面进行全局重复数据删除,以减少远程访问。它还能将类似的应用程序迁移到相同的 NUMA 节点上,从而实现无需远程访问的本地重复数据删除。我们还进一步使 AdaptMD 易于配置,以应对各种部署场景。实验表明,AdaptMD 可以像全局重复数据删除一样节省大量内存,同时实现与本地重复数据删除类似的访问性能。
{"title":"AdaptMD: Balancing Space and Performance in NUMA Architectures With Adaptive Memory Deduplication","authors":"Lulu Yao;Yongkun Li;Patrick P. C. Lee;Xiaoyang Wang;Yinlong Xu","doi":"10.1109/TC.2024.3375592","DOIUrl":"10.1109/TC.2024.3375592","url":null,"abstract":"Memory deduplication effectively relieves the memory space bottleneck by removing duplicate pages, especially in virtualized systems in which virtual machines run the same OS and similar applications. However, due to the non-uniform access latencies in NUMA architectures, memory deduplication poses a trade-off between memory savings and access performance: global deduplication across NUMA nodes realizes high memory savings, but leads to frequent cross-node remote access after deduplication and results in performance degradations. In contrast, local deduplication avoids remote access, but limits deduplication effectiveness. We design AdaptMD, an adaptive memory deduplication system that addresses the space-performance trade-off in NUMA architectures. AdaptMD leverages hotness awareness to globally deduplicate only cold pages to reduce remote access. It also migrates similar applications to the same NUMA node to allow local deduplication without remote access. We further make AdaptMD readily configurable to address various deployment scenarios. Experiments show that AdaptMD achieves high memory savings as in global deduplication, while achieving similar access performance as in local deduplication.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 6","pages":"1588-1602"},"PeriodicalIF":3.7,"publicationDate":"2024-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140154876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward Finding S-Box Circuits With Optimal Multiplicative Complexity 寻找具有最佳乘法复杂性的 S-Box 电路
IF 3.6 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-10 DOI: 10.1109/TC.2024.3398507
Yongjin Jeon;Seungjun Baek;Jongsung Kim
In this paper, we present a new method to find S-box circuits with optimal multiplicative complexity (MC), i.e., MC-optimal S-box circuits. We provide new observations for efficiently constructing circuits and computing MC, combined with a popular pathfinding algorithm named A*. In our search, the A* algorithm outputs a path of length MC, corresponding to an MC-optimal circuit. Based on an in-depth analysis of the process of computing MC, we enable the A* algorithm to function within our graph to investigate a wider range of S-boxes than existing methods such as the SAT-solver-based tool [1] and LIGHTER [2]. We provide implementable MC-optimal circuits for all the quadratic 5-bit bijective S-boxes and existing 5-bit almost-perfect nonlinear (APN) S-boxes. Furthermore, we present MC-optimal circuits for 6-bit S-boxes such as Sarkar Gold, Sarkar Quadratic, and some quadratic permutations. Finally, we theoretically demonstrate new lower bounds for the MCs of S-boxes, providing tighter bounds for the MCs of AES and MISTY S-boxes than previously known. This study complements previous results on MC-optimal S-box circuits and is intended to provide further insight into this field.
在本文中,我们提出了一种寻找具有最优乘法复杂度(MC)的 S-box 电路(即 MC-最优 S-box 电路)的新方法。我们为高效构建电路和计算 MC 提供了新的观测方法,并结合了一种名为 A* 的流行寻路算法。在我们的搜索中,A* 算法会输出一条长度为 MC 的路径,与 MC 最佳电路相对应。基于对 MC 计算过程的深入分析,与基于 SAT 求解器的工具 [1] 和 LIGHTER [2] 等现有方法相比,我们使 A* 算法在我们的图中能够研究更广泛的 S 框。我们为所有二次 5 位双射 S-box 和现有的 5 位几乎完全非线性 (APN) S-box 提供了可实现的 MC 最佳电路。此外,我们还提出了 6 位 S-box 的 MC 最佳电路,如 Sarkar Gold、Sarkar Quadratic 和一些二次排列。最后,我们从理论上证明了 S-box 的 MC 的新下限,为 AES 和 MISTY S-box 的 MC 提供了比以前已知的更严格的下限。这项研究补充了之前关于 MC 最佳 S-box 电路的结果,旨在为这一领域提供更深入的见解。
{"title":"Toward Finding S-Box Circuits With Optimal Multiplicative Complexity","authors":"Yongjin Jeon;Seungjun Baek;Jongsung Kim","doi":"10.1109/TC.2024.3398507","DOIUrl":"10.1109/TC.2024.3398507","url":null,"abstract":"In this paper, we present a new method to find S-box circuits with optimal multiplicative complexity (MC), i.e., MC-optimal S-box circuits. We provide new observations for efficiently constructing circuits and computing MC, combined with a popular pathfinding algorithm named A*. In our search, the A* algorithm outputs a path of length MC, corresponding to an MC-optimal circuit. Based on an in-depth analysis of the process of computing MC, we enable the A* algorithm to function within our graph to investigate a wider range of S-boxes than existing methods such as the SAT-solver-based tool \u0000<xref>[1]</xref>\u0000 and \u0000<monospace>LIGHTER</monospace>\u0000 \u0000<xref>[2]</xref>\u0000. We provide implementable MC-optimal circuits for all the quadratic 5-bit bijective S-boxes and existing 5-bit almost-perfect nonlinear (APN) S-boxes. Furthermore, we present MC-optimal circuits for 6-bit S-boxes such as Sarkar Gold, Sarkar Quadratic, and some quadratic permutations. Finally, we theoretically demonstrate new lower bounds for the MCs of S-boxes, providing tighter bounds for the MCs of \u0000<monospace>AES</monospace>\u0000 and \u0000<monospace>MISTY</monospace>\u0000 S-boxes than previously known. This study complements previous results on MC-optimal S-box circuits and is intended to provide further insight into this field.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 8","pages":"2036-2050"},"PeriodicalIF":3.6,"publicationDate":"2024-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140933829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CoDA: A Co-Design Framework for Versatile and Efficient Attention Accelerators CoDA:多功能高效注意力加速器的协同设计框架
IF 3.6 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-09 DOI: 10.1109/TC.2024.3398488
Wenjie Li;Aokun Hu;Ningyi Xu;Guanghui He
As a primary component of Transformers, attention mechanism suffers from quadratic computational complexity. To achieve efficient implementations, its hardware accelerator designs have aroused great research interest. However, most existing accelerators only support a single type of application and a single type of attention, making it difficult to meet the demands of diverse application scenarios. Additionally, they mainly focus on the dynamic pruning of attention matrices, which requires the deployment of pre-processing units, thereby reducing overall hardware efficiency. This paper presents CoDA which is an algorithm, dataflow and architecture co-design framework for versatile and efficient attention accelerators. The designed accelerator supports both NLP and CV applications, and can be configured into the mode supporting low-rank attention or low-rank plus sparse attention. We apply algorithmic transformations to low-rank attention to significantly reduce computational complexity. To prevent an increase in storage overhead resulting from the proposed algorithmic transformations, we carefully design the dataflows and adopt a block-wise fashion. Down-scaling softmax is further supported by architecture and dataflow co-design. Moreover, we propose a softmax sharing strategy to reduce the area cost. Our experiment results demonstrate that the proposed accelerator outperforms the state-of-the-art designs in terms of throughput, area efficiency and energy efficiency.
作为变形金刚的主要组成部分,注意力机制的计算复杂度高达四倍。为了实现高效实施,其硬件加速器设计引起了极大的研究兴趣。然而,现有的加速器大多只支持单一类型的应用和单一类型的注意力,难以满足多样化应用场景的需求。此外,它们主要关注注意力矩阵的动态剪枝,这需要部署预处理单元,从而降低了整体硬件效率。本文提出的 CoDA 是一个算法、数据流和架构协同设计框架,用于设计多功能、高效的注意力加速器。所设计的加速器支持 NLP 和 CV 应用,可配置为支持低秩注意力或低秩加稀疏注意力的模式。我们对低阶注意力进行了算法转换,以显著降低计算复杂度。为防止算法转换导致存储开销增加,我们精心设计了数据流,并采用了分块方式。架构和数据流的协同设计进一步支持了软最大值的缩减。此外,我们还提出了一种软最大共享策略,以降低面积成本。实验结果表明,所提出的加速器在吞吐量、面积效率和能效方面都优于最先进的设计。
{"title":"CoDA: A Co-Design Framework for Versatile and Efficient Attention Accelerators","authors":"Wenjie Li;Aokun Hu;Ningyi Xu;Guanghui He","doi":"10.1109/TC.2024.3398488","DOIUrl":"10.1109/TC.2024.3398488","url":null,"abstract":"As a primary component of Transformers, attention mechanism suffers from quadratic computational complexity. To achieve efficient implementations, its hardware accelerator designs have aroused great research interest. However, most existing accelerators only support a single type of application and a single type of attention, making it difficult to meet the demands of diverse application scenarios. Additionally, they mainly focus on the dynamic pruning of attention matrices, which requires the deployment of pre-processing units, thereby reducing overall hardware efficiency. This paper presents CoDA which is an algorithm, dataflow and architecture co-design framework for versatile and efficient attention accelerators. The designed accelerator supports both NLP and CV applications, and can be configured into the mode supporting low-rank attention or low-rank plus sparse attention. We apply algorithmic transformations to low-rank attention to significantly reduce computational complexity. To prevent an increase in storage overhead resulting from the proposed algorithmic transformations, we carefully design the dataflows and adopt a block-wise fashion. Down-scaling softmax is further supported by architecture and dataflow co-design. Moreover, we propose a softmax sharing strategy to reduce the area cost. Our experiment results demonstrate that the proposed accelerator outperforms the state-of-the-art designs in terms of throughput, area efficiency and energy efficiency.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 8","pages":"1924-1938"},"PeriodicalIF":3.6,"publicationDate":"2024-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140933994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Cost-Effective and Robust Packaging in Multi-Leader BFT Blockchain Systems 在多引线 BFT 区块链系统中实现经济高效的稳健封装
IF 3.6 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-09 DOI: 10.1109/TC.2024.3398510
Xiulong Liu;Zhiyuan Zheng;Wenbin Wang;Hao Xu;Fengjun Xiao;Keqiu Li
In Byzantine fault-tolerant (BFT) systems, maintaining consistency amidst malicious replicas is crucial, especially for blockchain systems. Recent innovations in this field have integrated multiple leaders into the BFT consensus mechanism to boost scalability and efficiency. However, the existing approaches often lead to excessive consumption of storage, bandwidth, and CPU resources due to redundant transactions. And the attempting to mitigate resource wastage inadvertently reduces resilience against Byzantine failures. To this end, we propose PeterHofe, an innovative ring-based approach for collaborative transaction processing. PeterHofe focuses on balancing resource utilization and minimizing the influence of Byzantine leaders, thereby enhancing transaction processing speed and overall system reliability. PeterHofe innovates by partitioning the transaction hash space into various buckets and creating a complex mappings between these buckets and the replicas, effectively reducing the control of Byzantine replicas. In developing PeterHofe, we concentrate on three primary objectives: 1) the creation of a permutation-based ring structure that enhances resistance to Byzantine censorship, backed by thorough mathematical proofs and analyses; 2) the development of a Prophecy-Implementation mechanism aimed at minimizing transaction replication while scrutinizing potential malicious activities; 3) to ensure the applicability of our proposed method across various types of multi-leader BFT consensus protocols, we have developed an additional asynchronous protocol to ensure consistent application of the packaging strategy. We have implemented PeterHofe using the latest significant frameworks, Narwhal and Tusk, and our empirical results affirm its capability to simultaneously minimize resource waste and bolster system robustness. Specifically, PeterHofe demonstrates efficiency in resource utilization, achieving a 20-fold reduction of resource waste when compared to the Random-based Strategy. When against the advanced Hash-based Partitioning Strategy, it reduces malicious transaction control by at least 66$%$, leading to up to 75$%$ lower latency. In scenarios of high traffic, our approach significantly outperforms existing strategies in throughput. Against the Random-based Strategy, it achieves a 6.11$%$ increase, and when compared to the Hash-based Partitioning Strategy, the improvement is 20$%$.
在拜占庭容错(BFT)系统中,在恶意复制中保持一致性至关重要,对于区块链系统尤其如此。该领域的最新创新将多个领导者整合到 BFT 共识机制中,以提高可扩展性和效率。然而,现有方法往往会因冗余交易而导致存储、带宽和 CPU 资源的过度消耗。为了减少资源浪费,我们无意中降低了对拜占庭故障的恢复能力。为此,我们提出了 PeterHofe,一种用于协作事务处理的基于环的创新方法。PeterHofe 专注于平衡资源利用率,尽量减少拜占庭领导者的影响,从而提高事务处理速度和整体系统可靠性。PeterHofe 的创新之处在于将事务散列空间划分为不同的桶,并在这些桶和副本之间创建复杂的映射,从而有效减少拜占庭副本的控制。在开发 PeterHofe 的过程中,我们专注于三个主要目标:1)创建一种基于置换的环状结构,以增强对拜占庭审查的抵抗能力,并辅以全面的数学证明和分析;2)开发一种预言实施机制,旨在最大限度地减少交易复制,同时审查潜在的恶意活动;3)为确保我们提出的方法适用于各种类型的多领导 BFT 共识协议,我们开发了一种额外的异步协议,以确保打包策略的一致应用。我们使用最新的重要框架 Narwhal 和 Tusk 实现了 PeterHofe,我们的实证结果肯定了它同时减少资源浪费和增强系统鲁棒性的能力。具体来说,PeterHofe 展示了资源利用效率,与基于随机的策略相比,资源浪费减少了 20 倍。与先进的基于哈希的分区策略相比,它至少减少了 66% 的恶意事务控制,从而降低了 75% 的延迟。在高流量场景中,我们的方法在吞吐量方面明显优于现有策略。与基于随机的策略相比,我们的方法提高了 6.11%;与基于哈希的分区策略相比,我们的方法提高了 20%。
{"title":"Towards Cost-Effective and Robust Packaging in Multi-Leader BFT Blockchain Systems","authors":"Xiulong Liu;Zhiyuan Zheng;Wenbin Wang;Hao Xu;Fengjun Xiao;Keqiu Li","doi":"10.1109/TC.2024.3398510","DOIUrl":"10.1109/TC.2024.3398510","url":null,"abstract":"In Byzantine fault-tolerant (BFT) systems, maintaining consistency amidst malicious replicas is crucial, especially for blockchain systems. Recent innovations in this field have integrated multiple leaders into the BFT consensus mechanism to boost scalability and efficiency. However, the existing approaches often lead to excessive consumption of storage, bandwidth, and CPU resources due to redundant transactions. And the attempting to mitigate resource wastage inadvertently reduces resilience against Byzantine failures. To this end, we propose PeterHofe, an innovative ring-based approach for collaborative transaction processing. PeterHofe focuses on balancing resource utilization and minimizing the influence of Byzantine leaders, thereby enhancing transaction processing speed and overall system reliability. PeterHofe innovates by partitioning the transaction hash space into various buckets and creating a complex mappings between these buckets and the replicas, effectively reducing the control of Byzantine replicas. In developing PeterHofe, we concentrate on three primary objectives: 1) the creation of a permutation-based ring structure that enhances resistance to Byzantine censorship, backed by thorough mathematical proofs and analyses; 2) the development of a Prophecy-Implementation mechanism aimed at minimizing transaction replication while scrutinizing potential malicious activities; 3) to ensure the applicability of our proposed method across various types of multi-leader BFT consensus protocols, we have developed an additional asynchronous protocol to ensure consistent application of the packaging strategy. We have implemented PeterHofe using the latest significant frameworks, Narwhal and Tusk, and our empirical results affirm its capability to simultaneously minimize resource waste and bolster system robustness. Specifically, PeterHofe demonstrates efficiency in resource utilization, achieving a 20-fold reduction of resource waste when compared to the Random-based Strategy. When against the advanced Hash-based Partitioning Strategy, it reduces malicious transaction control by at least 66\u0000<inline-formula><tex-math>$%$</tex-math></inline-formula>\u0000, leading to up to 75\u0000<inline-formula><tex-math>$%$</tex-math></inline-formula>\u0000 lower latency. In scenarios of high traffic, our approach significantly outperforms existing strategies in throughput. Against the Random-based Strategy, it achieves a 6.11\u0000<inline-formula><tex-math>$%$</tex-math></inline-formula>\u0000 increase, and when compared to the Hash-based Partitioning Strategy, the improvement is 20\u0000<inline-formula><tex-math>$%$</tex-math></inline-formula>\u0000.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 11","pages":"2590-2604"},"PeriodicalIF":3.6,"publicationDate":"2024-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140933762","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Neural Network Reliability: Insights From Hardware/Software Collaboration With Neuron Vulnerability Quantization 提高神经网络可靠性:神经元漏洞量化的硬件/软件合作启示
IF 3.6 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-09 DOI: 10.1109/TC.2024.3398492
Jing Wang;Jinbin Zhu;Xin Fu;Di Zang;Keyao Li;Weigong Zhang
Ensuring the reliability of deep neural networks (DNNs) is paramount in safety-critical applications. Although introducing supplementary fault-tolerant mechanisms can augment the reliability of DNNs, an efficiency tradeoff may be introduced. This study reveals the inherent fault tolerance of neural networks, where individual neurons exhibit varying degrees of fault tolerance, by thoroughly exploring the structural attributes of DNNs. We thereby develop a hardware/software collaborative method that guarantees the reliability of DNNs while minimizing performance degradation. We introduce the neuron vulnerability factor (NVF) to quantify the susceptibility to soft errors. We propose two efficient methods that leverage the NVF to minimize the negative effects of soft errors on neurons. First, we present a novel computational scheduling scheme. By prioritizing error-prone neurons, the expedited completion of their computations is facilitated to mitigate the risk of neural computing errors that arise from soft errors without sacrificing efficiency. Second, we propose the NVF-guided heterogeneous memory system. We employ variable-strength error-correcting codes and tailor their error-correction mechanisms to the vulnerability profile of specific neurons to ensure a highly targeted approach for error mitigation. Our experimental results demonstrate that the proposed scheme enhances the neural network accuracy by 18% on average, while significantly reducing the fault-tolerance overhead.
在安全关键型应用中,确保深度神经网络(DNN)的可靠性至关重要。虽然引入辅助容错机制可以增强 DNN 的可靠性,但可能会带来效率上的折衷。本研究通过深入探讨 DNN 的结构属性,揭示了神经网络固有的容错性,即单个神经元表现出不同程度的容错性。因此,我们开发了一种硬件/软件协作方法,既能保证 DNN 的可靠性,又能最大限度地减少性能下降。我们引入了神经元易损性因子(NVF)来量化对软错误的易感性。我们提出了两种有效的方法,利用 NVF 将软错误对神经元的负面影响降至最低。首先,我们提出了一种新颖的计算调度方案。通过对容易出错的神经元进行优先排序,加速完成其计算,从而在不牺牲效率的前提下,降低软错误导致的神经计算错误风险。其次,我们提出了 NVF 引导的异构存储系统。我们采用强度可变的纠错码,并根据特定神经元的脆弱性特征定制纠错机制,以确保采用极具针对性的方法来缓解错误。我们的实验结果表明,所提出的方案平均提高了 18% 的神经网络准确性,同时显著降低了容错开销。
{"title":"Enhancing Neural Network Reliability: Insights From Hardware/Software Collaboration With Neuron Vulnerability Quantization","authors":"Jing Wang;Jinbin Zhu;Xin Fu;Di Zang;Keyao Li;Weigong Zhang","doi":"10.1109/TC.2024.3398492","DOIUrl":"10.1109/TC.2024.3398492","url":null,"abstract":"Ensuring the reliability of deep neural networks (DNNs) is paramount in safety-critical applications. Although introducing supplementary fault-tolerant mechanisms can augment the reliability of DNNs, an efficiency tradeoff may be introduced. This study reveals the inherent fault tolerance of neural networks, where individual neurons exhibit varying degrees of fault tolerance, by thoroughly exploring the structural attributes of DNNs. We thereby develop a hardware/software collaborative method that guarantees the reliability of DNNs while minimizing performance degradation. We introduce the neuron vulnerability factor (NVF) to quantify the susceptibility to soft errors. We propose two efficient methods that leverage the NVF to minimize the negative effects of soft errors on neurons. First, we present a novel computational scheduling scheme. By prioritizing error-prone neurons, the expedited completion of their computations is facilitated to mitigate the risk of neural computing errors that arise from soft errors without sacrificing efficiency. Second, we propose the NVF-guided heterogeneous memory system. We employ variable-strength error-correcting codes and tailor their error-correction mechanisms to the vulnerability profile of specific neurons to ensure a highly targeted approach for error mitigation. Our experimental results demonstrate that the proposed scheme enhances the neural network accuracy by 18% on average, while significantly reducing the fault-tolerance overhead.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 8","pages":"1953-1966"},"PeriodicalIF":3.6,"publicationDate":"2024-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140934198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uniformity and Independence of H3 Hash Functions for Bloom Filters 布鲁姆过滤器 H3 哈希函数的统一性和独立性
IF 3.6 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-09 DOI: 10.1109/TC.2024.3398426
Furkan Koltuk;Ece Güran Schmidt
In this paper, we investigate the effects of violating the conditions of hash function uniformity and/or independence on the false positive probability of Bloom Filters (BF). To this end, we focus on hash functions of the H3 family with a partitioned memory organization for fast hardware implementations of BFs. We first introduce a dependence metric that quantifies hash function uniformity and independence. We then state and prove the necessary and sufficient conditions on the BF parameters for constructing uniform and independent hash functions. Finally, we derive an analytical expression for the exact false positive probability of a BF with hash functions that are not necessarily uniform or independent. We verify our expression with a hardware test bench and explore the effects of losing uniformity and independence through an experimental study that systematically sweeps different dependence metric values and numbers of hash functions. We demonstrate the effects of violating hash function uniformity and independence on the stated target false positive probability for selected previous works in the literature. As an important finding, we show that uniformity of individual hash functions is essential, whereas limited dependencies between hash functions can be tolerated without a negative effect on the false positive probability.
在本文中,我们研究了违反散列函数一致性和/或独立性条件对 Bloom Filters(BF)误报概率的影响。为此,我们将重点放在具有分区内存组织的 H3 系列散列函数上,以实现布隆过滤器的快速硬件实现。我们首先介绍一种量化哈希函数一致性和独立性的依赖性度量。然后,我们阐述并证明了构建统一和独立哈希函数的 BF 参数的必要条件和充分条件。最后,我们推导出一种分析表达式,用于计算散列函数不一定均匀或独立的 BF 的精确误报概率。我们用硬件测试台验证了我们的表达式,并通过系统地扫描不同依赖性度量值和散列函数数量的实验研究,探索了失去统一性和独立性的影响。我们展示了违反哈希函数统一性和独立性对文献中选定的先前作品的既定目标误报概率的影响。作为一项重要发现,我们表明单个哈希函数的一致性至关重要,而哈希函数之间的有限依赖性是可以容忍的,不会对误报概率产生负面影响。
{"title":"Uniformity and Independence of H3 Hash Functions for Bloom Filters","authors":"Furkan Koltuk;Ece Güran Schmidt","doi":"10.1109/TC.2024.3398426","DOIUrl":"10.1109/TC.2024.3398426","url":null,"abstract":"In this paper, we investigate the effects of violating the conditions of hash function uniformity and/or independence on the false positive probability of Bloom Filters (BF). To this end, we focus on hash functions of the H3 family with a partitioned memory organization for fast hardware implementations of BFs. We first introduce a dependence metric that quantifies hash function uniformity and independence. We then state and prove the necessary and sufficient conditions on the BF parameters for constructing uniform and independent hash functions. Finally, we derive an analytical expression for the exact false positive probability of a BF with hash functions that are not necessarily uniform or independent. We verify our expression with a hardware test bench and explore the effects of losing uniformity and independence through an experimental study that systematically sweeps different dependence metric values and numbers of hash functions. We demonstrate the effects of violating hash function uniformity and independence on the stated target false positive probability for selected previous works in the literature. As an important finding, we show that uniformity of individual hash functions is essential, whereas limited dependencies between hash functions can be tolerated without a negative effect on the false positive probability.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 8","pages":"1913-1923"},"PeriodicalIF":3.6,"publicationDate":"2024-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140933826","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ada-WL: An Adaptive Wear-Leveling Aware Data Migration Approach for Flexible SSD Array Scaling in Clusters Ada-WL:用于集群中灵活扩展固态硬盘阵列的自适应磨损水平感知数据迁移方法
IF 3.6 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-09 DOI: 10.1109/TC.2024.3398493
Yunfei Gu;Linhui Liu;Chentao Wu;Jie Li;Minyi Guo
Recently, the flash-based Solid State Drive (SSD) array has been widely implemented in real-world large-scale clusters. With the increasing number of users in upper-tier applications and the burst of Input/Output requests in this data explosive era, data centers need to continuously scale up to meet real-time data storage needs. However, the classical disk array scaling methods are designed based on HDDs, ignoring the wear leveling and garbage collection characteristics of SSD. This leads to penalties due to the vast lifetime gap between extended SSDs and the original in-use SSDs while scaling the SSD array, including extra triggered wear leveling I/O, latency in average response time, etc. To address these problems, we propose an Adaptive Wear-Leveling aware data migration approach for flexible SSD array scaling in clusters. It manages the interdisk wear leveling based on Model Reference Adaptive Control, which includes an SSD behavior emulator, Kalman filter estimator, and adaptive law. To demonstrate the effectiveness of this approach, we conducted several simulations and implementations on actual hardware. The evaluation results show that Ada-WL has the self-adaptability to optimize the wear leveling management parameters for various states of SSD arrays, diverse workloads, and scaling performed multiple times, significantly improving performance for SSD array scaling.
最近,基于闪存的固态硬盘(SSD)阵列在现实世界的大规模集群中得到了广泛应用。在这个数据爆炸的时代,随着上层应用用户数量的不断增加和输入/输出请求的激增,数据中心需要不断扩展以满足实时数据存储需求。然而,传统的磁盘阵列扩展方法是基于硬盘设计的,忽略了固态硬盘的磨损均衡和垃圾收集特性。这就导致在扩展固态硬盘阵列时,由于扩展固态硬盘与原始在用固态硬盘之间存在巨大的寿命差距而产生惩罚,包括额外触发的损耗均衡 I/O、平均响应时间延迟等。为解决这些问题,我们提出了一种自适应磨损水平感知数据迁移方法,用于在集群中灵活扩展固态硬盘阵列。该方法基于模型参考自适应控制来管理磁盘间损耗均衡,其中包括固态硬盘行为仿真器、卡尔曼滤波器估算器和自适应法则。为了证明这种方法的有效性,我们在实际硬件上进行了多次模拟和实施。评估结果表明,Ada-WL 具有自适应能力,能够针对固态硬盘阵列的各种状态、不同的工作负载和多次执行的扩展优化损耗均衡管理参数,显著提高了固态硬盘阵列扩展的性能。
{"title":"Ada-WL: An Adaptive Wear-Leveling Aware Data Migration Approach for Flexible SSD Array Scaling in Clusters","authors":"Yunfei Gu;Linhui Liu;Chentao Wu;Jie Li;Minyi Guo","doi":"10.1109/TC.2024.3398493","DOIUrl":"10.1109/TC.2024.3398493","url":null,"abstract":"Recently, the flash-based Solid State Drive (SSD) array has been widely implemented in real-world large-scale clusters. With the increasing number of users in upper-tier applications and the burst of Input/Output requests in this data explosive era, data centers need to continuously scale up to meet real-time data storage needs. However, the classical disk array scaling methods are designed based on HDDs, ignoring the wear leveling and garbage collection characteristics of SSD. This leads to penalties due to the vast lifetime gap between extended SSDs and the original in-use SSDs while scaling the SSD array, including extra triggered wear leveling I/O, latency in average response time, etc. To address these problems, we propose an Adaptive Wear-Leveling aware data migration approach for flexible SSD array scaling in clusters. It manages the interdisk wear leveling based on Model Reference Adaptive Control, which includes an SSD behavior emulator, Kalman filter estimator, and adaptive law. To demonstrate the effectiveness of this approach, we conducted several simulations and implementations on actual hardware. The evaluation results show that Ada-WL has the self-adaptability to optimize the wear leveling management parameters for various states of SSD arrays, diverse workloads, and scaling performed multiple times, significantly improving performance for SSD array scaling.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 8","pages":"1967-1982"},"PeriodicalIF":3.6,"publicationDate":"2024-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140933997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SimBU: Self-Similarity-Based Hybrid Binary-Unary Computing for Nonlinear Functions SimBU:基于自相似性的非线性函数混合二元统一计算
IF 3.6 2区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-03-09 DOI: 10.1109/TC.2024.3398512
Alireza Khataei;Gaurav Singh;Kia Bazargan
Unary computing is a relatively new method for implementing arbitrary nonlinear functions that uses unpacked thermometer number encoding, enabling much lower hardware costs. In its original form, unary computing provides no trade-off between accuracy and hardware cost. In this work, we propose a novel self-similarity-based method to optimize the previous hybrid binary-unary work and provide it with the trade-off between accuracy and hardware cost by introducing controlled levels of approximation. Looking for self-similarity between different parts of a function allows us to implement a very small subset of core unique subfunctions and derive the rest of the subfunctions from this core using simple linear transformations. We compare our method to previous works such as FloPoCo-LUT (lookup table), HBU (hybrid binary-unary) and FloPoCo-PPA (piecewise polynomial approximation) on several 8–12-bit nonlinear functions including Log, Exp, Sigmoid, GELU, Sin, and Sqr, which are frequently used in neural networks and image processing applications. The area $times$ delay hardware cost of our method is on average 32%–60% better than previous methods in both exact and approximate implementations. We also extend our method to multivariate nonlinear functions and show on average 78%–92% improvement over previous work.
一元计算是实现任意非线性函数的一种相对较新的方法,它使用未打包的温度计数字编码,使硬件成本大大降低。在其原始形式中,一元计算无法在精度和硬件成本之间做出权衡。在这项工作中,我们提出了一种基于自相似性的新方法,以优化之前的二元-一元混合工作,并通过引入可控的近似程度,在精度和硬件成本之间进行权衡。通过寻找函数不同部分之间的自相似性,我们可以实现极小的核心独特子函数子集,并通过简单的线性变换从该核心导出其余子函数。我们将我们的方法与以前的方法进行了比较,如 FloPoCo-LUT(查找表)、HBU(混合二元-一元)和 FloPoCo-PPA(分片多项式逼近),它们适用于多个 8-12 位非线性函数,包括 Log、Exp、Sigmoid、GELU、Sin 和 Sqr,这些函数在神经网络和图像处理应用中经常使用。在精确和近似实现方面,我们的方法的延迟硬件成本平均比以前的方法高出 32%-60%。我们还将我们的方法扩展到多变量非线性函数,结果显示比以前的工作平均提高了 78%-92% 。
{"title":"SimBU: Self-Similarity-Based Hybrid Binary-Unary Computing for Nonlinear Functions","authors":"Alireza Khataei;Gaurav Singh;Kia Bazargan","doi":"10.1109/TC.2024.3398512","DOIUrl":"10.1109/TC.2024.3398512","url":null,"abstract":"Unary computing is a relatively new method for implementing arbitrary nonlinear functions that uses unpacked thermometer number encoding, enabling much lower hardware costs. In its original form, unary computing provides no trade-off between accuracy and hardware cost. In this work, we propose a novel self-similarity-based method to optimize the previous hybrid binary-unary work and provide it with the trade-off between accuracy and hardware cost by introducing controlled levels of approximation. Looking for self-similarity between different parts of a function allows us to implement a very small subset of core unique subfunctions and derive the rest of the subfunctions from this core using simple linear transformations. We compare our method to previous works such as FloPoCo-LUT (lookup table), HBU (hybrid binary-unary) and FloPoCo-PPA (piecewise polynomial approximation) on several 8–12-bit nonlinear functions including Log, Exp, Sigmoid, GELU, Sin, and Sqr, which are frequently used in neural networks and image processing applications. The area \u0000<inline-formula><tex-math>$times$</tex-math></inline-formula>\u0000 delay hardware cost of our method is on average 32%–60% better than previous methods in both exact and approximate implementations. We also extend our method to multivariate nonlinear functions and show on average 78%–92% improvement over previous work.","PeriodicalId":13087,"journal":{"name":"IEEE Transactions on Computers","volume":"73 9","pages":"2192-2205"},"PeriodicalIF":3.6,"publicationDate":"2024-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140933756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Computers
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1