首页 > 最新文献

Microprocessors and Microsystems最新文献

英文 中文
Retraction notice to the articles published in the Special issue Smart Agri from “Microprocessors and Microsystems” 关于 "微处理器与微系统 "特刊《智能农业》中发表文章的撤稿通知
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-28 DOI: 10.1016/j.micpro.2024.105038
{"title":"Retraction notice to the articles published in the Special issue Smart Agri from “Microprocessors and Microsystems”","authors":"","doi":"10.1016/j.micpro.2024.105038","DOIUrl":"https://doi.org/10.1016/j.micpro.2024.105038","url":null,"abstract":"","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"106 ","pages":"Article 105038"},"PeriodicalIF":2.6,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0141933124000334/pdfft?md5=6970ffb236c663c09734849d8c298072&pid=1-s2.0-S0141933124000334-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139993273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A light-weight neuromorphic controlling clock gating based multi-core cryptography platform 基于时钟门控的轻量级神经形态多核加密平台
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-26 DOI: 10.1016/j.micpro.2024.105040
Pham-Khoi Dong , Khanh N. Dang , Duy-Anh Nguyen , Xuan-Tu Tran

While speeding up cryptography tasks can be accomplished by using a multi-core architecture to parallelize computation, one of the major challenges is optimizing power consumption. In principle, depending on the computation workload, individual cores can be turned off to save power during operation. However, too few active cores may lead to computational bottlenecks. In this work, we propose a novel platform named Spike-MCryptCores: a low-power multi-core AES platform with a neuromorphic controller. The proposed Spike-MCryptCores platform is composed of multiple AES cores, each core is equipped with a clock-gating scheme for reducing its power consumption while being idle. To optimize the power consumption of the whole platform, we use a neuromorphic controller. Therefore, a comprehensive framework to generate a data set, train the neural network, and produce hardware configuration for the Spiking Neural Network (SNN), a brain-inspired computing paradigm, is also presented in this paper. Moreover, Spike-MCryptCores integrates the hardware SNN inside its architecture to support low-cost and low-latency adaptations. The results show that implemented SNN controller occupies only 2.3 % of the overall area cost while providing the ability to reduce power consumption significantly. The lightweight SNN controller model is trained and tested with up to 95 % accuracy. The maximum difference between the predicted number of cores and the ideal one from the label is one unit only. Under 24 test scenarios, a SNN controller with clock-gating helps Spike-MCryptCores reducing the power consumption by 48.6 % on the average; by 67 % for the best-case scenario, and by 39 % for the worst-case scenario.

虽然利用多核架构并行计算可以加快密码学任务的速度,但其中一个主要挑战是优化功耗。原则上,根据计算工作量,可以在运行期间关闭单个内核以节省功耗。然而,过少的活动内核可能会导致计算瓶颈。在这项工作中,我们提出了一种名为 Spike-MCryptCores 的新型平台:一种带有神经形态控制器的低功耗多核 AES 平台。所提出的 Spike-MCryptCores 平台由多个 AES 内核组成,每个内核都配备了时钟门方案,以降低空闲时的功耗。为了优化整个平台的功耗,我们使用了神经形态控制器。因此,本文还提出了一个综合框架,用于生成数据集、训练神经网络,以及为尖峰神经网络(SNN)(一种大脑启发计算范例)生成硬件配置。此外,Spike-MCryptCores 还在其架构中集成了硬件 SNN,以支持低成本和低延迟的自适应。研究结果表明,实现的 SNN 控制器仅占整体面积成本的 2.3%,同时还能显著降低功耗。轻量级 SNN 控制器模型经过训练和测试,准确率高达 95%。根据标签预测的内核数与理想内核数之间的最大差异仅为一个单位。在 24 种测试场景下,带有时钟门的 SNN 控制器帮助 Spike-MCryptCores 平均降低了 48.6% 的功耗;在最佳场景下降低了 67%,在最差场景下降低了 39%。
{"title":"A light-weight neuromorphic controlling clock gating based multi-core cryptography platform","authors":"Pham-Khoi Dong ,&nbsp;Khanh N. Dang ,&nbsp;Duy-Anh Nguyen ,&nbsp;Xuan-Tu Tran","doi":"10.1016/j.micpro.2024.105040","DOIUrl":"10.1016/j.micpro.2024.105040","url":null,"abstract":"<div><p>While speeding up cryptography tasks can be accomplished by using a multi-core architecture to parallelize computation, one of the major challenges is optimizing power consumption. In principle, depending on the computation workload, individual cores can be turned off to save power during operation. However, too few active cores may lead to computational bottlenecks. In this work, we propose a novel platform named Spike-MCryptCores: a low-power multi-core AES platform with a neuromorphic controller. The proposed Spike-MCryptCores platform is composed of multiple AES cores, each core is equipped with a clock-gating scheme for reducing its power consumption while being idle. To optimize the power consumption of the whole platform, we use a neuromorphic controller. Therefore, a comprehensive framework to generate a data set, train the neural network, and produce hardware configuration for the Spiking Neural Network (SNN), a brain-inspired computing paradigm, is also presented in this paper. Moreover, Spike-MCryptCores integrates the hardware SNN inside its architecture to support low-cost and low-latency adaptations. The results show that implemented SNN controller occupies only 2.3 % of the overall area cost while providing the ability to reduce power consumption significantly. The lightweight SNN controller model is trained and tested with up to 95 % accuracy. The maximum difference between the predicted number of cores and the ideal one from the label is one unit only. Under 24 test scenarios, a SNN controller with clock-gating helps Spike-MCryptCores reducing the power consumption by 48.6 % on the average; by 67 % for the best-case scenario, and by 39 % for the worst-case scenario.</p></div>","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"106 ","pages":"Article 105040"},"PeriodicalIF":2.6,"publicationDate":"2024-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139987897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MOSAIC: Maximizing ResOurce Sharing in Behavioral Application SpecIfic ProCessors MOSAIC:最大限度地共享行为应用程序模拟处理器中的资源
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-20 DOI: 10.1016/j.micpro.2024.105039
Qilin Si, Benjamin Carrion Schafer

This work presents a method that can quickly determine which hardware accelerators (HWaccs) should be mapped together onto an Application-Specific Instruction Set Processor (ASIP), such that the resources shared among them are maximized. This work in particular targets HWaccs generated from untimed behavioral descriptions for High-Level Synthesis (HLS). Although HLS is a single process synthesis method, our approach is able to force resource sharing among the HWaccs by combining their behavioral descriptions together into a single description based on their potential to share resources. These shared resources include functional units (FUs) like multipliers, adders, and dividers, and also registers. In particular, our proposed flow leads up to 48% in area savings and on average 30%. Because an exhaustive enumeration of all possible combinations can lead to long runtimes, we propose a fast heuristic that leads to comparable results (only 6% worse on average), while being much faster (on average 500×).

这项工作提出了一种方法,可以快速确定哪些硬件加速器(HWaccs)应一起映射到特定应用指令集处理器(ASIP)上,从而使它们之间共享的资源最大化。这项工作特别针对从用于高级合成(HLS)的未计时行为描述中生成的 HWaccs。虽然 HLS 是一种单进程合成方法,但我们的方法能够根据 HWaccs 的资源共享潜力,将它们的行为描述合并为单一描述,从而强制它们共享资源。这些共享资源包括乘法器、加法器、除法器等功能单元(FU)以及寄存器。特别是,我们提出的流程可节省高达 48% 的面积,平均节省 30%。由于穷举所有可能的组合会导致较长的运行时间,因此我们提出了一种快速启发式方法,其结果与之相当(平均只差 6%),同时速度更快(平均 500 倍)。
{"title":"MOSAIC: Maximizing ResOurce Sharing in Behavioral Application SpecIfic ProCessors","authors":"Qilin Si,&nbsp;Benjamin Carrion Schafer","doi":"10.1016/j.micpro.2024.105039","DOIUrl":"10.1016/j.micpro.2024.105039","url":null,"abstract":"<div><p>This work presents a method that can quickly determine which hardware accelerators (HWaccs) should be mapped together onto an Application-Specific Instruction Set Processor (ASIP), such that the resources shared among them are maximized. This work in particular targets HWaccs generated from untimed behavioral descriptions for High-Level Synthesis (HLS). Although HLS is a single process synthesis method, our approach is able to force resource sharing among the HWaccs by combining their behavioral descriptions together into a single description based on their potential to share resources. These shared resources include functional units (FUs) like multipliers, adders, and dividers, and also registers. In particular, our proposed flow leads up to 48% in area savings and on average 30%. Because an exhaustive enumeration of all possible combinations can lead to long runtimes, we propose a fast heuristic that leads to comparable results (only 6% worse on average), while being much faster (on average 500<span><math><mo>×</mo></math></span>).</p></div>","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"106 ","pages":"Article 105039"},"PeriodicalIF":2.6,"publicationDate":"2024-02-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139919853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Real-time P-SFA hardware implementation of Deep Neural Networks using FPGA 使用 FPGA 的深度神经网络实时 P-SFA 硬件实现
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-17 DOI: 10.1016/j.micpro.2024.105037
Nour Elshahawy , Sandy A. Wasif , Maggie Mashaly , Eman Azab

Machine Learning (ML) algorithms, specifically Artificial Neural Networks (ANNs), have proved their effectiveness in solving complex problems in many different applications and multiple fields. This paper focuses on optimizing the activation function (AF) block of the NN hardware architecture. The AF block used is based on a probability-based sigmoid function approximation block (P-SFA) combined with a novel real-time probability module (PRT) that calculates the probability of the input data. The proposed NN design aims to use the least amount of hardware resources and area while maintaining a high recognition accuracy. The proposed AF module in this work consists of two P-SFA blocks and the PRT component. The architecture proposed for implementing NNs is evaluated on Field Programmable Gate Arrays (FPGAs). The proposed design has achieved a recognition accuracy of 97.84 % on a 6-layer Deep Neural Network (DNN) for the MNIST dataset and a recognition accuracy of 88.58% on a 6-layer DNN for the FMNIST dataset. The proposed AF module has a total area of 1136 LUTs and 327 FFs, a logical critical path delay of 8.853 ns. The power consumption of the P-SFA block is 6 mW and the PRT block is 5 mW.

机器学习(ML)算法,特别是人工神经网络(ANN),已证明能有效解决许多不同应用和多个领域的复杂问题。本文的重点是优化 NN 硬件架构中的激活函数 (AF) 模块。所使用的激活函数块基于基于概率的 sigmoid 函数近似块 (P-SFA),并与计算输入数据概率的新型实时概率模块 (PRT) 相结合。拟议的 NN 设计旨在使用最少的硬件资源和面积,同时保持较高的识别准确率。这项工作中提出的自动指纹识别模块由两个 P-SFA 模块和 PRT 组件组成。在现场可编程门阵列(FPGA)上评估了为实现 NN 而提出的架构。在 MNIST 数据集上,拟议设计的 6 层深度神经网络(DNN)的识别准确率达到 97.84%,在 FMNIST 数据集上,拟议设计的 6 层深度神经网络的识别准确率达到 88.58%。拟议的自动指纹识别模块的总面积为 1136 个 LUT 和 327 个 FF,逻辑关键路径延迟为 8.853 ns。P-SFA 模块的功耗为 6 mW,PRT 模块的功耗为 5 mW。
{"title":"A Real-time P-SFA hardware implementation of Deep Neural Networks using FPGA","authors":"Nour Elshahawy ,&nbsp;Sandy A. Wasif ,&nbsp;Maggie Mashaly ,&nbsp;Eman Azab","doi":"10.1016/j.micpro.2024.105037","DOIUrl":"10.1016/j.micpro.2024.105037","url":null,"abstract":"<div><p>Machine Learning (ML) algorithms, specifically Artificial Neural Networks (ANNs), have proved their effectiveness in solving complex problems in many different applications and multiple fields. This paper focuses on optimizing the activation function (AF) block of the NN hardware architecture. The AF block used is based on a probability-based sigmoid function approximation block (P-SFA) combined with a novel real-time probability module (PRT) that calculates the probability of the input data. The proposed NN design aims to use the least amount of hardware resources and area while maintaining a high recognition accuracy. The proposed AF module in this work consists of two P-SFA blocks and the PRT component. The architecture proposed for implementing NNs is evaluated on Field Programmable Gate Arrays (FPGAs). The proposed design has achieved a recognition accuracy of 97.84 % on a 6-layer Deep Neural Network (DNN) for the MNIST dataset and a recognition accuracy of 88.58% on a 6-layer DNN for the FMNIST dataset. The proposed AF module has a total area of 1136 LUTs and 327 FFs, a logical critical path delay of 8.853 ns. The power consumption of the P-SFA block is 6 mW and the PRT block is 5 mW.</p></div>","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"106 ","pages":"Article 105037"},"PeriodicalIF":2.6,"publicationDate":"2024-02-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139919876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design and evaluation of low power and area efficient approximate Booth multipliers for error tolerant applications 设计和评估用于容错应用的低功耗、高效面积近似布斯乘法器
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-17 DOI: 10.1016/j.micpro.2024.105036
Vishal Gundavarapu , P. Gowtham , A. Anita Angeline , P. Sasipriya

Approximate computing is an innovative design methodology to reduce the design complexity with an improvement in power efficiency, performance and area by compromising on the requirement of accuracy. In this paper, 8-bit approximate Booth multipliers have been proposed based on the approximate Radix-4 modified Booth encoding algorithm and approximate compressors for partial product accumulation to produce the final products are proposed. Two approximate Probability Based Booth Encoders (PBBE-1 and PBBE-2) have been proposed and used in the Booth multipliers. Error parameters have been measured and compared with the existing approximate booth multipliers. Exact booth multiplier of novel design existing in the literature has also been implemented for comparison purpose. The proposed approximate multipliers are then used in applications like image multiplication and IIR bi-quad filtering to prove their performance. Simulation results prove that the proposed booth multipliers outperform the existing approximate booth multipliers in terms of power and area with better accuracy. Synthesis results prove that the proposed Multiplier 6 was found to be the most efficient with a 56 % power consumption improvement and a 47 % area improvement when compared to the exact multiplier. All the simulations are carried out using Cadence® Genus with 180 nm CMOS process technology.

近似计算是一种创新的设计方法,它可以降低设计复杂度,提高能效、性能和面积,同时又不影响精度要求。本文提出了基于近似 Radix-4 改良 Booth 编码算法的 8 位近似 Booth 乘法器,并提出了用于部分乘积累加以生成最终乘积的近似压缩器。我们提出了两种基于概率的近似 Booth 编码器(PBBE-1 和 PBBE-2),并将其用于 Booth 乘法器中。对误差参数进行了测量,并与现有的近似展位乘法器进行了比较。为了进行比较,还实现了文献中现有的新颖设计的精确展位乘法器。然后,将所提出的近似乘法器用于图像乘法和 IIR 双四滤波等应用中,以证明其性能。仿真结果证明,所提出的近似亭乘法器在功耗和面积方面优于现有的近似亭乘法器,而且精度更高。合成结果证明,与精确乘法器相比,建议的乘法器 6 功耗降低了 56%,面积缩小了 47%,是最高效的乘法器。所有仿真均采用 180 纳米 CMOS 工艺技术的 Cadence® Genus 进行。
{"title":"Design and evaluation of low power and area efficient approximate Booth multipliers for error tolerant applications","authors":"Vishal Gundavarapu ,&nbsp;P. Gowtham ,&nbsp;A. Anita Angeline ,&nbsp;P. Sasipriya","doi":"10.1016/j.micpro.2024.105036","DOIUrl":"https://doi.org/10.1016/j.micpro.2024.105036","url":null,"abstract":"<div><p>Approximate computing is an innovative design methodology to reduce the design complexity with an improvement in power efficiency, performance and area by compromising on the requirement of accuracy. In this paper, 8-bit approximate Booth multipliers have been proposed based on the approximate Radix-4 modified Booth encoding algorithm and approximate compressors for partial product accumulation to produce the final products are proposed. Two approximate Probability Based Booth Encoders (PBBE-1 and PBBE-2) have been proposed and used in the Booth multipliers. Error parameters have been measured and compared with the existing approximate booth multipliers. Exact booth multiplier of novel design existing in the literature has also been implemented for comparison purpose. The proposed approximate multipliers are then used in applications like image multiplication and IIR bi-quad filtering to prove their performance. Simulation results prove that the proposed booth multipliers outperform the existing approximate booth multipliers in terms of power and area with better accuracy. Synthesis results prove that the proposed Multiplier 6 was found to be the most efficient with a 56 % power consumption improvement and a 47 % area improvement when compared to the exact multiplier. All the simulations are carried out using Cadence® Genus with 180 nm CMOS process technology.</p></div>","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"106 ","pages":"Article 105036"},"PeriodicalIF":2.6,"publicationDate":"2024-02-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139908537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Retraction notice to “FPGA implementation of low power and high speed image edge detection algorithm” [Microprocessors and Microsystems 75, 2020, 103053] 关于 "低功耗和高速图像边缘检测算法的 FPGA 实现 "的撤稿通知 [Microprocessors and Microsystems 75, 2020, 103053]
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-14 DOI: 10.1016/j.micpro.2024.105024
R. Menaka , R. Janarthanan , K. Deeba
{"title":"Retraction notice to “FPGA implementation of low power and high speed image edge detection algorithm” [Microprocessors and Microsystems 75, 2020, 103053]","authors":"R. Menaka ,&nbsp;R. Janarthanan ,&nbsp;K. Deeba","doi":"10.1016/j.micpro.2024.105024","DOIUrl":"https://doi.org/10.1016/j.micpro.2024.105024","url":null,"abstract":"","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"105 ","pages":"Article 105024"},"PeriodicalIF":2.6,"publicationDate":"2024-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S014193312400019X/pdfft?md5=50252cb4ed13eb15d928f39b6a36d8df&pid=1-s2.0-S014193312400019X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139731480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Retraction notice to 'A machine learning based IoT for providing an intrusion detection system for security' [Microprocessors and Microsystems 82 (2021) 103741] 基于机器学习的物联网为安全提供入侵检测系统 "的撤稿通知 [Microprocessors and Microsystems 82 (2021) 103741]
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-14 DOI: 10.1016/j.micpro.2024.105028
Dhanke Jyoti Atul , R. Kamalraj , G. Ramesh , K. Sakthidasan Sankaran , Sudhir Sharma , Syed Khasim
{"title":"Retraction notice to 'A machine learning based IoT for providing an intrusion detection system for security' [Microprocessors and Microsystems 82 (2021) 103741]","authors":"Dhanke Jyoti Atul ,&nbsp;R. Kamalraj ,&nbsp;G. Ramesh ,&nbsp;K. Sakthidasan Sankaran ,&nbsp;Sudhir Sharma ,&nbsp;Syed Khasim","doi":"10.1016/j.micpro.2024.105028","DOIUrl":"https://doi.org/10.1016/j.micpro.2024.105028","url":null,"abstract":"","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"105 ","pages":"Article 105028"},"PeriodicalIF":2.6,"publicationDate":"2024-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0141933124000231/pdfft?md5=2324ddffca814a8e7e1d9a95bf25bcd5&pid=1-s2.0-S0141933124000231-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139731510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Retraction notice to ‘Virtual garden landscape planning based on FPGA and GIS platform’ [Microprocessors and Microsystems 79 (2020) 103314] 基于 FPGA 和 GIS 平台的虚拟园林景观规划"[Microprocessors and Microsystems 79 (2020) 103314] 撤稿通知
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-14 DOI: 10.1016/j.micpro.2024.105025
Xiaoxia Bai
{"title":"Retraction notice to ‘Virtual garden landscape planning based on FPGA and GIS platform’ [Microprocessors and Microsystems 79 (2020) 103314]","authors":"Xiaoxia Bai","doi":"10.1016/j.micpro.2024.105025","DOIUrl":"https://doi.org/10.1016/j.micpro.2024.105025","url":null,"abstract":"","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"105 ","pages":"Article 105025"},"PeriodicalIF":2.6,"publicationDate":"2024-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0141933124000206/pdfft?md5=3ca6695ae7be1714954a1ad923ad1ca0&pid=1-s2.0-S0141933124000206-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139731508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Retraction notice to 'An Improved Dynamic Process Neural Network Prediction Model Identification Method' [Microprocessors and Microsystems 80 (2021) 103573] 改进的动态过程神经网络预测模型识别方法》的撤稿通知 [Microprocessors and Microsystems 80 (2021) 103573]
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-14 DOI: 10.1016/j.micpro.2024.105026
Shuran Lyu , Peng Liu , Lu Liu , Shuqi Ma , Tao Wang
{"title":"Retraction notice to 'An Improved Dynamic Process Neural Network Prediction Model Identification Method' [Microprocessors and Microsystems 80 (2021) 103573]","authors":"Shuran Lyu ,&nbsp;Peng Liu ,&nbsp;Lu Liu ,&nbsp;Shuqi Ma ,&nbsp;Tao Wang","doi":"10.1016/j.micpro.2024.105026","DOIUrl":"https://doi.org/10.1016/j.micpro.2024.105026","url":null,"abstract":"","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"105 ","pages":"Article 105026"},"PeriodicalIF":2.6,"publicationDate":"2024-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0141933124000218/pdfft?md5=6f2dd19803dcd14e81cb81974d56d475&pid=1-s2.0-S0141933124000218-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139731509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Flip-and-Patch: A fault-tolerant technique for on-chip memories of CNN accelerators at low supply voltage 翻转与补丁:低电源电压下 CNN 加速器片上存储器的容错技术
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2024-02-10 DOI: 10.1016/j.micpro.2024.105023
Yamilka Toca-Díaz , Reynier Hernández Palacios , Rubén Gran Tejero , Alejandro Valero

Aggressively reducing the supply voltage (Vdd) below the safe threshold voltage (Vmin) can effectively lead to significant energy savings in digital circuits. However, operating at such low supply voltages poses challenges due to a high occurrence of permanent faults resulting from manufacturing process variations in current technology nodes.

This work addresses the impact of permanent faults on the accuracy of a Convolutional Neural Network (CNN) inference accelerator using on-chip activation memories supplied at low Vdd below Vmin. Based on a characterization study of fault patterns, this paper proposes two low-cost microarchitectural techniques, namely Flip-and-Patch, which maintain the original accuracy of CNN applications even in the presence of a high number of faults caused by operating at Vdd<Vmin. Unlike existing techniques, Flip-and-Patch remains transparent to the programmer and does not rely on application characteristics, making it easily applicable to real CNN accelerators.

Experimental results show that Flip-and-Patch ensures the original CNN accuracy with a minimal impact on system performance (less than 0.05% for every application), while achieving average energy savings of 10.5% and 46.6% in activation memories compared to a conventional accelerator operating at safe and nominal supply voltages, respectively. Compared to the state-of-the-art ThUnderVolt technique, which dynamically adjusts the supply voltage at run time and discarding any energy overhead for such an approach, the average energy savings are by 3.2%.

积极地将电源电压(Vdd)降低到安全阈值电压(Vmin)以下,可以有效地为数字电路节省大量能源。然而,由于当前技术节点的制造工艺差异导致永久性故障的高发生率,在如此低的电源电压下工作带来了挑战。本研究探讨了永久性故障对卷积神经网络(CNN)推理加速器准确性的影响,该加速器使用的是在低于 Vmin 的低 Vdd 下供电的片上激活存储器。基于对故障模式的特性研究,本文提出了两种低成本微体系结构技术,即翻转和修补技术,即使在 Vdd<Vmin 工作时出现大量故障,也能保持 CNN 应用的原始精度。实验结果表明,Flip-and-Patch 可确保 CNN 的原始精度,对系统性能的影响极小(每个应用均小于 0.05%),同时与在安全和额定电源电压下运行的传统加速器相比,激活内存的平均节能率分别为 10.5%和 46.6%。最先进的 ThUnderVolt 技术可在运行时动态调整电源电压,而这种方法不需要任何能源开销,与之相比,平均节能 3.2%。
{"title":"Flip-and-Patch: A fault-tolerant technique for on-chip memories of CNN accelerators at low supply voltage","authors":"Yamilka Toca-Díaz ,&nbsp;Reynier Hernández Palacios ,&nbsp;Rubén Gran Tejero ,&nbsp;Alejandro Valero","doi":"10.1016/j.micpro.2024.105023","DOIUrl":"https://doi.org/10.1016/j.micpro.2024.105023","url":null,"abstract":"<div><p>Aggressively reducing the supply voltage (<span><math><msub><mrow><mi>V</mi></mrow><mrow><mi>d</mi><mi>d</mi></mrow></msub></math></span>) below the safe threshold voltage (<span><math><msub><mrow><mi>V</mi></mrow><mrow><mi>m</mi><mi>i</mi><mi>n</mi></mrow></msub></math></span>) can effectively lead to significant energy savings in digital circuits. However, operating at such low supply voltages poses challenges due to a high occurrence of permanent faults resulting from manufacturing process variations in current technology nodes.</p><p>This work addresses the impact of permanent faults on the accuracy of a Convolutional Neural Network (CNN) inference accelerator using on-chip activation memories supplied at low <span><math><msub><mrow><mi>V</mi></mrow><mrow><mi>d</mi><mi>d</mi></mrow></msub></math></span> below <span><math><msub><mrow><mi>V</mi></mrow><mrow><mi>m</mi><mi>i</mi><mi>n</mi></mrow></msub></math></span>. Based on a characterization study of fault patterns, this paper proposes two low-cost microarchitectural techniques, namely Flip-and-Patch, which maintain the original accuracy of CNN applications even in the presence of a high number of faults caused by operating at <span><math><mrow><msub><mrow><mi>V</mi></mrow><mrow><mi>d</mi><mi>d</mi></mrow></msub><mo>&lt;</mo><msub><mrow><mi>V</mi></mrow><mrow><mi>m</mi><mi>i</mi><mi>n</mi></mrow></msub></mrow></math></span>. Unlike existing techniques, Flip-and-Patch remains transparent to the programmer and does not rely on application characteristics, making it easily applicable to real CNN accelerators.</p><p>Experimental results show that Flip-and-Patch ensures the original CNN accuracy with a minimal impact on system performance (less than 0.05% for every application), while achieving average energy savings of 10.5% and 46.6% in activation memories compared to a conventional accelerator operating at safe and nominal supply voltages, respectively. Compared to the state-of-the-art ThUnderVolt technique, which dynamically adjusts the supply voltage at run time and discarding any energy overhead for such an approach, the average energy savings are by 3.2%.</p></div>","PeriodicalId":49815,"journal":{"name":"Microprocessors and Microsystems","volume":"106 ","pages":"Article 105023"},"PeriodicalIF":2.6,"publicationDate":"2024-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0141933124000188/pdfft?md5=9ccb11570430c8c998f414582a020757&pid=1-s2.0-S0141933124000188-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139726887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Microprocessors and Microsystems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1