首页 > 最新文献

IEEE Transactions on Emerging Topics in Computing最新文献

英文 中文
Always on Voting: A Framework for Repetitive Voting on the Blockchain 永远投票:区块链上的重复投票框架
IF 5.9 2区 计算机科学 Q1 Computer Science Pub Date : 2023-09-22 DOI: 10.1109/TETC.2023.3315748
Sarad Venugopalan;Ivana Stančíková;Ivan Homoliak
Elections repeat commonly after a fixed time interval, ranging from months to years. This results in limitations on governance since elected candidates or policies are difficult to remove before the next elections, if needed, and allowed by the corresponding law. Participants may decide (through a public deliberation) to change their choices but have no opportunity to vote for these choices before the next elections. Another issue is the peak-end effect, where the judgment of voters is based on how they felt a short time before the elections. To address these issues, we propose Always on Voting (AoV) – a repetitive voting framework that allows participants to vote and change elected candidates or policies without waiting for the next elections. Participants are permitted to privately change their vote at any point in time, while the effect of their change is manifested at the end of each epoch, whose duration is shorter than the time between two main elections. To thwart the problem of peak-end effect in epochs, the ends of epochs are randomized and made unpredictable, while preserved within soft bounds. These goals are achieved using the synergy between a Bitcoin puzzle oracle, verifiable delay function, and smart contracts.
选举通常在固定的时间间隔后重复进行,间隔时间从数月到数年不等。这就造成了对治理的限制,因为当选的候选人或政策很难在下次选举前(如果需要)被撤换,而且相应的法律也允许这样做。参与者可以(通过公开讨论)决定改变他们的选择,但没有机会在下次选举前对这些选择进行投票。另一个问题是峰末效应,即选民的判断是基于选举前不久的感受。为了解决这些问题,我们提出了 "随时投票"(Always on Voting,AoV)--一种重复投票框架,允许参与者投票并改变当选的候选人或政策,而无需等待下一次选举。参与者可以在任何时间点私下更改投票,而更改的效果会在每个纪元结束时体现出来,每个纪元的持续时间比两次主要选举之间的时间短。为了避免历时峰值效应的问题,历时的结束时间是随机的,不可预测,同时保持在软约束范围内。这些目标是通过比特币谜题甲骨文、可验证延迟函数和智能合约之间的协同作用来实现的。
{"title":"Always on Voting: A Framework for Repetitive Voting on the Blockchain","authors":"Sarad Venugopalan;Ivana Stančíková;Ivan Homoliak","doi":"10.1109/TETC.2023.3315748","DOIUrl":"https://doi.org/10.1109/TETC.2023.3315748","url":null,"abstract":"Elections repeat commonly after a fixed time interval, ranging from months to years. This results in limitations on governance since elected candidates or policies are difficult to remove before the next elections, if needed, and allowed by the corresponding law. Participants may decide (through a public deliberation) to change their choices but have no opportunity to vote for these choices before the next elections. Another issue is the peak-end effect, where the judgment of voters is based on how they felt a short time before the elections. To address these issues, we propose Always on Voting (AoV) – a repetitive voting framework that allows participants to vote and change elected candidates or policies without waiting for the next elections. Participants are permitted to privately change their vote at any point in time, while the effect of their change is manifested at the end of each epoch, whose duration is shorter than the time between two main elections. To thwart the problem of peak-end effect in epochs, the ends of epochs are randomized and made unpredictable, while preserved within soft bounds. These goals are achieved using the synergy between a Bitcoin puzzle oracle, verifiable delay function, and smart contracts.","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.9,"publicationDate":"2023-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138558040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Hardware/Software Co-Design With ADC-Less In-Memory Computing Hardware for Spiking Neural Networks 针对尖峰神经网络的无 ADC 内存计算硬件的硬件/软件协同设计
IF 5.9 2区 计算机科学 Q1 Computer Science Pub Date : 2023-09-22 DOI: 10.1109/TETC.2023.3316121
Marco Paul E. Apolinario;Adarsh Kumar Kosta;Utkarsh Saxena;Kaushik Roy
Spiking Neural Networks (SNNs) are bio-plausible models that hold great potential for realizing energy-efficient implementations of sequential tasks on resource-constrained edge devices. However, commercial edge platforms based on standard GPUs are not optimized to deploy SNNs, resulting in high energy and latency. While analog In-Memory Computing (IMC) platforms can serve as energy-efficient inference engines, they are accursed by the immense energy, latency, and area requirements of high-precision ADCs (HP-ADC), overshadowing the benefits of in-memory computations. We propose a hardware/software co-design methodology to deploy SNNs into an ADC-Less IMC architecture using sense-amplifiers as 1-bit ADCs replacing conventional HP-ADCs and alleviating the above issues. Our proposed framework incurs minimal accuracy degradation by performing hardware-aware training and is able to scale beyond simple image classification tasks to more complex sequential regression tasks. Experiments on complex tasks of optical flow estimation and gesture recognition show that progressively increasing the hardware awareness during SNN training allows the model to adapt and learn the errors due to the non-idealities associated with ADC-Less IMC. Also, the proposed ADC-Less IMC offers significant energy and latency improvements, $2-7times$ and $8.9-24.6times$, respectively, depending on the SNN model and the workload, compared to HP-ADC IMC.
尖峰神经网络(SNN)是一种生物仿真模型,在资源受限的边缘设备上实现高能效的连续任务实施方面具有巨大潜力。然而,基于标准 GPU 的商用边缘平台并未针对部署 SNN 进行优化,从而导致高能耗和高延迟。虽然模拟内存计算(IMC)平台可以作为高能效推理引擎,但高精度模数转换器(HP-ADC)对能耗、延迟和面积的要求使其不堪重负,从而掩盖了内存计算的优势。我们提出了一种硬件/软件协同设计方法,将 SNN 部署到无 ADC IMC 架构中,使用感测放大器作为 1 位 ADC,取代传统的 HP-ADC,从而缓解上述问题。我们提出的框架通过执行硬件感知训练,将准确性降低到最低程度,并且能够从简单的图像分类任务扩展到更复杂的连续回归任务。光流估计和手势识别等复杂任务的实验表明,在 SNN 训练过程中逐步提高硬件感知能力,可使模型适应并学习与无 ADC IMC 相关的非理想性所造成的错误。此外,与 HP-ADC IMC 相比,根据 SNN 模型和工作负载的不同,拟议的无 ADC IMC 能显著改善能耗和延迟,分别为 2-7 美元/次和 8.9-24.6 美元/次。
{"title":"Hardware/Software Co-Design With ADC-Less In-Memory Computing Hardware for Spiking Neural Networks","authors":"Marco Paul E. Apolinario;Adarsh Kumar Kosta;Utkarsh Saxena;Kaushik Roy","doi":"10.1109/TETC.2023.3316121","DOIUrl":"10.1109/TETC.2023.3316121","url":null,"abstract":"Spiking Neural Networks (SNNs) are bio-plausible models that hold great potential for realizing energy-efficient implementations of sequential tasks on resource-constrained edge devices. However, commercial edge platforms based on standard GPUs are not optimized to deploy SNNs, resulting in high energy and latency. While analog In-Memory Computing (IMC) platforms can serve as energy-efficient inference engines, they are accursed by the immense energy, latency, and area requirements of high-precision ADCs (HP-ADC), overshadowing the benefits of in-memory computations. We propose a hardware/software co-design methodology to deploy SNNs into an ADC-Less IMC architecture using sense-amplifiers as 1-bit ADCs replacing conventional HP-ADCs and alleviating the above issues. Our proposed framework incurs minimal accuracy degradation by performing hardware-aware training and is able to scale beyond simple image classification tasks to more complex sequential regression tasks. Experiments on complex tasks of optical flow estimation and gesture recognition show that progressively increasing the hardware awareness during SNN training allows the model to adapt and learn the errors due to the non-idealities associated with ADC-Less IMC. Also, the proposed ADC-Less IMC offers significant energy and latency improvements, \u0000<inline-formula><tex-math>$2-7times$</tex-math></inline-formula>\u0000 and \u0000<inline-formula><tex-math>$8.9-24.6times$</tex-math></inline-formula>\u0000, respectively, depending on the SNN model and the workload, compared to HP-ADC IMC.","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.9,"publicationDate":"2023-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135599176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Concept Stability Entropy: A Novel Group Cohesion Measure in Social Networks 概念稳定性熵:社交网络中一种新的群体凝聚力测量方法
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-09-20 DOI: 10.1109/TETC.2023.3315335
Fei Hao;Jie Gao;Yaguang Lin;Yulei Wu;Jiaxing Shang
Group cohesion is regarded as a central group property across both social psychology and sociology. It facilities the understanding of the organizational behavior of users, and in turn guides the users to work well together in order to achieve goals within a social network. Therefore, group cohesion assessment is a crucial research issue for social network analysis. Group cohesion is often viewed as network density in the current state-of-the-art. Due to the advantages of characterizing the cohesion of a network with concept stability, this article presents a novel group cohesion measure, called concept stability entropy inspired by Shannon Entropy. Particularly, the scale of concept stability entropy is investigated. Considering the dynamic nature of social networks, an incremental algorithm for concept stability entropy computation is devised. In addition, we explore the correlation between concept stability entropy and other related metrics, i.e., network density, average degree, and average clustering coefficient. Extensive experimental results first validate that the concept stability entropy falls into the range of $[0, log(k)]$ ($k$ is the number of formal concepts), and then demonstrate that the concept stability entropy has a positive correlation with the average degree and a negative correlation with the network density and average clustering coefficient. Practically, a case study on the COVID-2019 virus network is conducted for illustrating the usefulness of our proposed group cohesion assessment approach.
在社会心理学和社会学中,群体凝聚力都被视为群体的核心属性。它有助于理解用户的组织行为,进而引导用户在社会网络中为实现目标而通力合作。因此,群体凝聚力评估是社会网络分析的一个重要研究课题。在当前最先进的技术中,群体凝聚力通常被视为网络密度。鉴于用概念稳定性来表征网络凝聚力的优势,本文受香农熵(Shannon Entropy)的启发,提出了一种新的群体凝聚力测量方法--概念稳定性熵。本文特别研究了概念稳定熵的尺度。考虑到社交网络的动态性质,我们设计了一种概念稳定熵计算的增量算法。此外,我们还探讨了概念稳定熵与其他相关指标(即网络密度、平均度和平均聚类系数)之间的相关性。大量实验结果首先验证了概念稳定熵的范围为 $[0,log(k)]$($k$ 为正式概念的个数),然后证明了概念稳定熵与平均度呈正相关,而与网络密度和平均聚类系数呈负相关。在实践中,我们对 COVID-2019 病毒网络进行了案例研究,以说明我们提出的群体凝聚力评估方法的实用性。
{"title":"Concept Stability Entropy: A Novel Group Cohesion Measure in Social Networks","authors":"Fei Hao;Jie Gao;Yaguang Lin;Yulei Wu;Jiaxing Shang","doi":"10.1109/TETC.2023.3315335","DOIUrl":"10.1109/TETC.2023.3315335","url":null,"abstract":"Group cohesion is regarded as a central group property across both social psychology and sociology. It facilities the understanding of the organizational behavior of users, and in turn guides the users to work well together in order to achieve goals within a social network. Therefore, group cohesion assessment is a crucial research issue for social network analysis. Group cohesion is often viewed as network density in the current state-of-the-art. Due to the advantages of characterizing the cohesion of a network with concept stability, this article presents a novel group cohesion measure, called concept stability entropy inspired by Shannon Entropy. Particularly, the scale of concept stability entropy is investigated. Considering the dynamic nature of social networks, an incremental algorithm for concept stability entropy computation is devised. In addition, we explore the correlation between concept stability entropy and other related metrics, i.e., network density, average degree, and average clustering coefficient. Extensive experimental results first validate that the concept stability entropy falls into the range of \u0000<inline-formula><tex-math>$[0, log(k)]$</tex-math></inline-formula>\u0000 (\u0000<inline-formula><tex-math>$k$</tex-math></inline-formula>\u0000 is the number of formal concepts), and then demonstrate that the concept stability entropy has a positive correlation with the average degree and a negative correlation with the network density and average clustering coefficient. Practically, a case study on the COVID-2019 virus network is conducted for illustrating the usefulness of our proposed group cohesion assessment approach.","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2023-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135556035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deadline-Aware and Energy-Efficient Dynamic Task Mapping and Scheduling for Multicore Systems Based on Wireless Network-on-Chip 基于无线片上网络的多核系统的截止时间感知和高能效动态任务映射与调度
IF 5.9 2区 计算机科学 Q1 Computer Science Pub Date : 2023-09-20 DOI: 10.1109/TETC.2023.3315298
Abbas Dehghani;Sadegh Fadaei;Bahman Ravaei;Keyvan RahimiZadeh
Hybrid Wireless Network-on-Chip (HWNoC) architecture has been introduced as a promising communication infrastructure for multicore systems. HWNoC-based multicore systems encounter extremely dynamic application workloads that are submitted at run-time. Mapping and scheduling of these applications are critical for system performance, especially for real-time applications. The existing resource allocation approaches either ignore the use of wireless links in task allocation on cores or ignore the timing characteristic of tasks. In this paper, we propose a new deadline-aware and energy-efficient dynamic task mapping and scheduling approach for the HWNoC-based multicore system. By using of core utilization threshold and tasks laxity time, the proposed approach aims to minimize communication energy consumption and satisfy the deadline of the real-time applications tasks. Through cycle-accurate simulation, the performance of the proposed approach has been compared with state-of-the-art approaches in terms of communication energy consumption, deadline violation rate, communication latency, and runtime overhead. The experimental results confirmed that the proposed approach is a very competitive approach among the alternative approaches.
片上混合无线网络(HWNoC)架构作为一种有前途的多核系统通信基础设施已被引入。基于 HWNoC 的多核系统会遇到在运行时提交的极其动态的应用工作负载。这些应用的映射和调度对系统性能至关重要,尤其是对实时应用而言。现有的资源分配方法要么忽略了内核任务分配中无线链路的使用,要么忽略了任务的时序特性。在本文中,我们为基于 HWNoC 的多核系统提出了一种新的截止日期感知和高能效动态任务映射与调度方法。通过使用内核利用率阈值和任务松弛时间,该方法旨在最大限度地减少通信能耗,并满足实时应用任务的截止日期要求。通过周期精确仿真,从通信能耗、违反截止日期率、通信延迟和运行时开销等方面比较了所提方法与最先进方法的性能。实验结果证实,所提出的方法在其他方法中是一种非常有竞争力的方法。
{"title":"Deadline-Aware and Energy-Efficient Dynamic Task Mapping and Scheduling for Multicore Systems Based on Wireless Network-on-Chip","authors":"Abbas Dehghani;Sadegh Fadaei;Bahman Ravaei;Keyvan RahimiZadeh","doi":"10.1109/TETC.2023.3315298","DOIUrl":"10.1109/TETC.2023.3315298","url":null,"abstract":"Hybrid Wireless Network-on-Chip (HWNoC) architecture has been introduced as a promising communication infrastructure for multicore systems. HWNoC-based multicore systems encounter extremely dynamic application workloads that are submitted at run-time. Mapping and scheduling of these applications are critical for system performance, especially for real-time applications. The existing resource allocation approaches either ignore the use of wireless links in task allocation on cores or ignore the timing characteristic of tasks. In this paper, we propose a new deadline-aware and energy-efficient dynamic task mapping and scheduling approach for the HWNoC-based multicore system. By using of core utilization threshold and tasks laxity time, the proposed approach aims to minimize communication energy consumption and satisfy the deadline of the real-time applications tasks. Through cycle-accurate simulation, the performance of the proposed approach has been compared with state-of-the-art approaches in terms of communication energy consumption, deadline violation rate, communication latency, and runtime overhead. The experimental results confirmed that the proposed approach is a very competitive approach among the alternative approaches.","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.9,"publicationDate":"2023-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135555776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding Bulk-Bitwise Processing In-Memory Through Database Analytics 通过数据库分析了解内存中的批量比特处理
IF 5.9 2区 计算机科学 Q1 Computer Science Pub Date : 2023-09-19 DOI: 10.1109/TETC.2023.3315189
Ben Perach;Ronny Ronen;Benny Kimelfeld;Shahar Kvatinsky
Bulk-bitwise processing-in-memory (PIM), where large bitwise operations are performed in parallel by the memory array itself, is an emerging form of computation with the potential to mitigate the memory wall problem. This article examines the capabilities of bulk-bitwise PIM by constructing PIMDB, a fully-digital system based on memristive stateful logic, utilizing and focusing on in-memory bulk-bitwise operations, designed to accelerate a real-life workload: analytical processing of relational databases. We introduce a host processor programming model to support bulk-bitwise PIM in virtual memory, develop techniques to efficiently perform in-memory filtering and aggregation operations, and adapt the application data set into the memory. To understand bulk-bitwise PIM, we compare it to an equivalent in-memory database on the same host system. We show that bulk-bitwise PIM substantially lowers the number of required memory read operations, thus accelerating TPC-H filter operations by 1.6×–18× and full queries by 56×–608×, while reducing the energy consumption by 1.7×–18.6× and 0.81×–12× for these benchmarks, respectively. Our extensive evaluation uses the gem5 full-system simulation environment. The simulations also evaluate cell endurance, showing that the required endurance is within the range of existing endurance of RRAM devices.
内存中的批量位操作(PIM)是一种新兴的计算形式,它由内存阵列本身并行执行大型位操作,有望缓解内存墙问题。本文通过构建基于内存有状态逻辑的全数字系统 PIMDB,利用并专注于内存中的批量位操作,研究了批量位操作 PIM 的能力,该系统旨在加速现实生活中的工作负载:关系数据库的分析处理。我们引入了一种主机处理器编程模型,以支持虚拟内存中的批量位向 PIM,开发了高效执行内存过滤和聚合操作的技术,并将应用数据集调整到内存中。为了理解批量位向 PIM,我们将其与同一主机系统上的等效内存数据库进行了比较。我们的结果表明,bulk-bitwise PIM 大幅降低了所需内存读取操作的数量,从而将 TPC-H 筛选操作的速度提高了 1.6×-18×,将完整查询的速度提高了 56×-608×,同时将这些基准的能耗分别降低了 1.7×-18.6× 和 0.81×-12×。我们使用 gem5 全系统仿真环境进行了广泛的评估。仿真还评估了单元的耐用性,结果表明所需的耐用性在现有 RRAM 器件的耐用性范围之内。
{"title":"Understanding Bulk-Bitwise Processing In-Memory Through Database Analytics","authors":"Ben Perach;Ronny Ronen;Benny Kimelfeld;Shahar Kvatinsky","doi":"10.1109/TETC.2023.3315189","DOIUrl":"10.1109/TETC.2023.3315189","url":null,"abstract":"Bulk-bitwise processing-in-memory (PIM), where large bitwise operations are performed in parallel by the memory array itself, is an emerging form of computation with the potential to mitigate the memory wall problem. This article examines the capabilities of bulk-bitwise PIM by constructing PIMDB, a fully-digital system based on memristive stateful logic, utilizing and focusing on in-memory bulk-bitwise operations, designed to accelerate a real-life workload: analytical processing of relational databases. We introduce a host processor programming model to support bulk-bitwise PIM in virtual memory, develop techniques to efficiently perform in-memory filtering and aggregation operations, and adapt the application data set into the memory. To understand bulk-bitwise PIM, we compare it to an equivalent in-memory database on the same host system. We show that bulk-bitwise PIM substantially lowers the number of required memory read operations, thus accelerating TPC-H filter operations by 1.6×–18× and full queries by 56×–608×, while reducing the energy consumption by 1.7×–18.6× and 0.81×–12× for these benchmarks, respectively. Our extensive evaluation uses the gem5 full-system simulation environment. The simulations also evaluate cell endurance, showing that the required endurance is within the range of existing endurance of RRAM devices.","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.9,"publicationDate":"2023-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135551143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Guest Editorial Special Section on Applied Software Aging and Rejuvenation 应用软件老化与再生特邀编辑专区
IF 5.9 2区 计算机科学 Q1 Computer Science Pub Date : 2023-09-05 DOI: 10.1109/TETC.2023.3299150
Michael Grottke;Alberto Avritzer;Hironori Washizaki;Kishor Trivedi
Since the publication of the first paper on software aging and rejuvenation by Huang et al. in 1995 [1], considerable research has been devoted to this topic. It deals with the phenomenon that continuously-running software systems may show an increasing failure rate and/or a degrading performance, either because error conditions accumulate inside the running system or because the rate at which faults are activated and errors are propagated is positively correlated with system uptime. Software rejuvenation relates to techniques counteracting aging (for example, by regularly stopping and restarting the software) in order to remove aging effects and to proactively prevent failures from occurring.
自1995年Huang等人发表第一篇关于软件老化与返老返老的论文以来,这一主题得到了大量的研究。它处理的现象是,连续运行的软件系统可能会显示出越来越高的故障率和/或性能下降,这要么是因为错误条件在运行的系统中积累,要么是因为故障被激活和错误传播的速度与系统正常运行时间呈正相关。软件再生涉及对抗老化的技术(例如,通过定期停止和重新启动软件),以消除老化影响并主动防止故障发生。
{"title":"Guest Editorial Special Section on Applied Software Aging and Rejuvenation","authors":"Michael Grottke;Alberto Avritzer;Hironori Washizaki;Kishor Trivedi","doi":"10.1109/TETC.2023.3299150","DOIUrl":"10.1109/TETC.2023.3299150","url":null,"abstract":"Since the publication of the first paper on software aging and rejuvenation by Huang et al. in 1995 [1], considerable research has been devoted to this topic. It deals with the phenomenon that continuously-running software systems may show an increasing failure rate and/or a degrading performance, either because error conditions accumulate inside the running system or because the rate at which faults are activated and errors are propagated is positively correlated with system uptime. Software rejuvenation relates to techniques counteracting aging (for example, by regularly stopping and restarting the software) in order to remove aging effects and to proactively prevent failures from occurring.","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.9,"publicationDate":"2023-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10241255","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"62529112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Emerging Topics in Computing Information for Authors 面向作者的计算信息新兴主题IEEE汇刊
IF 5.9 2区 计算机科学 Q1 Computer Science Pub Date : 2023-09-05 DOI: 10.1109/TETC.2023.3300132
{"title":"IEEE Transactions on Emerging Topics in Computing Information for Authors","authors":"","doi":"10.1109/TETC.2023.3300132","DOIUrl":"10.1109/TETC.2023.3300132","url":null,"abstract":"","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.9,"publicationDate":"2023-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10241262","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135804629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Multiplier-Free RNS-Based CNN Accelerator Exploiting Bit-Level Sparsity 利用位级稀疏性的无乘法器 RNS 型 CNN 加速器
IF 5.9 2区 计算机科学 Q1 Computer Science Pub Date : 2023-08-10 DOI: 10.1109/TETC.2023.3301590
Vasilis Sakellariou;Vassilis Paliouras;Ioannis Kouretas;Hani Saleh;Thanos Stouraitis
In this work, a Residue Numbering System (RNS)-based Convolutional Neural Network (CNN) accelerator utilizing a multiplier-free distributed-arithmetic Processing Element (PE) is proposed. A method for maximizing the utilization of the arithmetic hardware resources is presented. It leads to an increase of the system's throughput, by exploiting bit-level sparsity within the weight vectors. The proposed PE design takes advantage of the properties of RNS and Canonical Signed Digit (CSD) encoding to achieve higher energy efficiency and effective processing rate, without requiring any compression mechanism or introducing any approximation. An extensive design space exploration for various parameters (RNS base, PE micro-architecture, encoding) using analytical models as well as experimental results from CNN benchmarks is conducted and the various trade-offs are analyzed. A complete end-to-end RNS accelerator is developed based on the proposed PE. The introduced accelerator is compared to traditional binary and RNS counterparts as well as to other state-of-the-art systems. Implementation results in a 22-nm process show that the proposed PE can lead to $1.85times$ and $1.54times$ more energy-efficient processing compared to binary and conventional RNS, respectively, with a $1.88times$ maximum increase of effective throughput for the employed benchmarks. Compared to a state-of-the-art, all-digital, RNS-based system, the proposed accelerator is $8.87times$ and $1.11times$ more energy- and area-efficient, respectively.
本研究提出了一种基于残差编码系统(RNS)的卷积神经网络(CNN)加速器,该加速器采用了无乘法器分布式算术处理元件(PE)。它提出了一种最大化算术硬件资源利用率的方法。通过利用权重向量中的位级稀疏性,该方法提高了系统的吞吐量。拟议的 PE 设计利用了 RNS 和 Canonical Signed Digit (CSD) 编码的特性,实现了更高的能效和有效处理率,而不需要任何压缩机制或引入任何近似值。利用分析模型和 CNN 基准的实验结果,对各种参数(RNS 基础、PE 微体系结构、编码)进行了广泛的设计空间探索,并对各种权衡进行了分析。基于所提出的 PE,开发了一个完整的端到端 RNS 加速器。将引入的加速器与传统的二进制和 RNS 对应系统以及其他最先进的系统进行了比较。在 22 纳米工艺中的实现结果表明,与二进制和传统 RNS 相比,所提出的 PE 可使能效处理分别提高 1.85 倍和 1.54 倍,所采用基准的有效吞吐量最大提高 1.88 倍。与最先进的全数字 RNS 系统相比,所提出的加速器的能效和面积效率分别提高了 8.87 倍和 1.11 倍。
{"title":"A Multiplier-Free RNS-Based CNN Accelerator Exploiting Bit-Level Sparsity","authors":"Vasilis Sakellariou;Vassilis Paliouras;Ioannis Kouretas;Hani Saleh;Thanos Stouraitis","doi":"10.1109/TETC.2023.3301590","DOIUrl":"10.1109/TETC.2023.3301590","url":null,"abstract":"In this work, a Residue Numbering System (RNS)-based Convolutional Neural Network (CNN) accelerator utilizing a multiplier-free distributed-arithmetic Processing Element (PE) is proposed. A method for maximizing the utilization of the arithmetic hardware resources is presented. It leads to an increase of the system's throughput, by exploiting bit-level sparsity within the weight vectors. The proposed PE design takes advantage of the properties of RNS and Canonical Signed Digit (CSD) encoding to achieve higher energy efficiency and effective processing rate, without requiring any compression mechanism or introducing any approximation. An extensive design space exploration for various parameters (RNS base, PE micro-architecture, encoding) using analytical models as well as experimental results from CNN benchmarks is conducted and the various trade-offs are analyzed. A complete end-to-end RNS accelerator is developed based on the proposed PE. The introduced accelerator is compared to traditional binary and RNS counterparts as well as to other state-of-the-art systems. Implementation results in a 22-nm process show that the proposed PE can lead to \u0000<inline-formula><tex-math>$1.85times$</tex-math></inline-formula>\u0000 and \u0000<inline-formula><tex-math>$1.54times$</tex-math></inline-formula>\u0000 more energy-efficient processing compared to binary and conventional RNS, respectively, with a \u0000<inline-formula><tex-math>$1.88times$</tex-math></inline-formula>\u0000 maximum increase of effective throughput for the employed benchmarks. Compared to a state-of-the-art, all-digital, RNS-based system, the proposed accelerator is \u0000<inline-formula><tex-math>$8.87times$</tex-math></inline-formula>\u0000 and \u0000<inline-formula><tex-math>$1.11times$</tex-math></inline-formula>\u0000 more energy- and area-efficient, respectively.","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.9,"publicationDate":"2023-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"62529168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Noise-Shaping Binary-to-Stochastic Converters for Reduced-Length Bit-Streams 用于缩短比特流长度的噪声整形二进制到随机转换器
IF 5.9 2区 计算机科学 Q1 Computer Science Pub Date : 2023-08-01 DOI: 10.1109/TETC.2023.3299516
Kleanthis Papachatzopoulos;Vassilis Paliouras
Stochastic computations have attracted significant attention for applications with moderate fixed-point accuracy requirements, as they offer minimal complexity. In these systems, a stochastic bit-stream encodes a data sample. The derived bit-stream is used for processing. The bit-stream length determines the computation latency for bit-serial implementations and hardware complexity for bit-parallel ones. Noise shaping is a feedback technique that moves the quantization noise outside the bandwidth of interest of a signal. This article proposes a technique that builds on noise shaping and reduces the length of the stochastic bit-stream required to achieve a specific Signal-to-Quantization-Noise Ratio (SQNR). The technique is realized by digital units that encode binary samples into stochastic streams, hereafter called as binary-to-stochastic converters. Furthermore, formulas are derived that relate the bit-stream length reduction to the signal bandwidth. First-order and second-order converters that implement the proposed technique are analyzed. Two architectures are introduced, distinguished by placing a stochastic converter either inside or outside of the noise-shaping loop. The proposed bit-stream length reduction is quantitatively compared to conventional binary-to-stochastic converters for the same signal quality level. Departing from conventional approaches, this article employs bit-stream lengths that are not a power of two, and proposes a modified stochastic-to-binary conversion scheme as a part of the proposed binary-to-stochastic converter. Particularly, SQNR gains of 29.8 dB and 42.1 dB are achieved for the first-order and second-order converters compared to the conventional converters for equal-length bit-streams and low signal bandwidth. The investigated converters are designed and synthesized at a 28-nm FDSOI technology for a range of bit widths.
对于定点精度要求适中的应用来说,随机计算因其复杂性最低而备受关注。在这些系统中,随机比特流对数据样本进行编码。导出的比特流用于处理。位流长度决定了位串行实现的计算延迟和位并行实现的硬件复杂度。噪声整形是一种反馈技术,可将量化噪声移至信号相关带宽之外。本文提出了一种以噪声整形为基础的技术,可减少实现特定信噪比(SQNR)所需的随机比特流长度。该技术通过将二进制采样编码为随机流的数字单元(以下称为二进制-随机转换器)来实现。此外,还推导出了比特流长度缩减与信号带宽相关的公式。分析了实现拟议技术的一阶和二阶转换器。通过将随机转换器置于噪声整形环路内部或外部,介绍了两种不同的架构。在信号质量水平相同的情况下,与传统的二进制到随机转换器进行了定量比较,发现所提出的比特流长度缩减效果更好。与传统方法不同的是,本文采用的比特流长度不是二的幂次,并提出了一种改进的随机到二进制转换方案,作为拟议的二进制到随机转换器的一部分。与传统转换器相比,一阶和二阶转换器在等长比特流和低信号带宽条件下的 SQNR 分别提高了 29.8 dB 和 42.1 dB。所研究的转换器采用 28 纳米 FDSOI 技术设计和合成,适用于各种位宽。
{"title":"Noise-Shaping Binary-to-Stochastic Converters for Reduced-Length Bit-Streams","authors":"Kleanthis Papachatzopoulos;Vassilis Paliouras","doi":"10.1109/TETC.2023.3299516","DOIUrl":"10.1109/TETC.2023.3299516","url":null,"abstract":"Stochastic computations have attracted significant attention for applications with moderate fixed-point accuracy requirements, as they offer minimal complexity. In these systems, a stochastic bit-stream encodes a data sample. The derived bit-stream is used for processing. The bit-stream length determines the computation latency for bit-serial implementations and hardware complexity for bit-parallel ones. Noise shaping is a feedback technique that moves the quantization noise outside the bandwidth of interest of a signal. This article proposes a technique that builds on noise shaping and reduces the length of the stochastic bit-stream required to achieve a specific Signal-to-Quantization-Noise Ratio (SQNR). The technique is realized by digital units that encode binary samples into stochastic streams, hereafter called as binary-to-stochastic converters. Furthermore, formulas are derived that relate the bit-stream length reduction to the signal bandwidth. First-order and second-order converters that implement the proposed technique are analyzed. Two architectures are introduced, distinguished by placing a stochastic converter either inside or outside of the noise-shaping loop. The proposed bit-stream length reduction is quantitatively compared to conventional binary-to-stochastic converters for the same signal quality level. Departing from conventional approaches, this article employs bit-stream lengths that are not a power of two, and proposes a modified stochastic-to-binary conversion scheme as a part of the proposed binary-to-stochastic converter. Particularly, SQNR gains of 29.8 dB and 42.1 dB are achieved for the first-order and second-order converters compared to the conventional converters for equal-length bit-streams and low signal bandwidth. The investigated converters are designed and synthesized at a 28-nm FDSOI technology for a range of bit widths.","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.9,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"62529127","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Edge-Cloud Collaboration Framework for Graph Processing in Smart Society 智能社会中图形处理的边缘-云协作框架
IF 5.9 2区 计算机科学 Q1 Computer Science Pub Date : 2023-07-24 DOI: 10.1109/TETC.2023.3297066
Jun Zhou;Masaaki Kondo
Due to the limitations of cloud computing on latency, bandwidth and data confidentiality, edge computing has emerged as a novel location-aware way to provide the capacity-constrained portable terminals with more processing capacity to improve the computing performance and quality of service (QoS) in several typical domains of the human activity in smart society, such as social networks, medical diagnosis, telecommunications, recommendation systems, internal threat detection, transportation, Internet of Things (IoT), etc. These application domains often manage a vast collection of entities with various relationships, which can be naturally represented by the graph data structure. Graph processing is a powerful tool to model and optimize complex problems where graph-based data is involved. In consideration of the relatively insufficient resource provisioning of the edge devices, in this article, for the first time to our knowledge, we propose a reliable edge-cloud collaboration framework that facilitates the graph primitives based on a lightweight interactive graph processing library (GPL), especially for shortest path search (SPS) operations as the demonstrative example. Two types of different practical cases are also presented to show the typical application scenarios of our graph processing strategy. Experimental evaluations indicate that the acceleration rate of performance can reach 6.87x via graph reduction, and less than 3% and 20% extra latency is required for much better user experiences for navigation and pandemic control, respectively, while the online security measures merely consume about 1% extra time of the overall data transmission. Our framework can efficiently execute the applications with considering of user-friendliness, low-latency response, interactions among edge devices, collaboration between edge and cloud, and privacy protection at an acceptable overhead.
由于云计算在延迟、带宽和数据保密性等方面的限制,边缘计算作为一种新颖的位置感知方式应运而生,它可以为容量受限的便携式终端提供更多处理能力,从而提高计算性能和服务质量(QoS),适用于智能社会中人类活动的多个典型领域,如社交网络、医疗诊断、电信、推荐系统、内部威胁检测、交通、物联网(IoT)等。这些应用领域通常管理着大量具有各种关系的实体集合,这些实体可以自然地用图数据结构来表示。图处理是对涉及图数据的复杂问题进行建模和优化的强大工具。考虑到边缘设备的资源供应相对不足,我们在本文中首次提出了一个可靠的边缘-云协作框架,该框架基于轻量级交互式图处理库(GPL),特别是以最短路径搜索(SPS)操作为例,促进了图原语的使用。此外,还介绍了两类不同的实际案例,以展示我们的图处理策略的典型应用场景。实验评估表明,通过图缩减,性能加速率可达 6.87 倍,导航和大流行病控制分别只需不到 3% 和 20% 的额外延迟,就能获得更好的用户体验,而在线安全措施只需消耗整体数据传输约 1% 的额外时间。考虑到用户友好性、低延迟响应、边缘设备之间的交互、边缘与云之间的协作以及隐私保护,我们的框架能以可接受的开销高效地执行应用程序。
{"title":"An Edge-Cloud Collaboration Framework for Graph Processing in Smart Society","authors":"Jun Zhou;Masaaki Kondo","doi":"10.1109/TETC.2023.3297066","DOIUrl":"10.1109/TETC.2023.3297066","url":null,"abstract":"Due to the limitations of cloud computing on latency, bandwidth and data confidentiality, edge computing has emerged as a novel location-aware way to provide the capacity-constrained portable terminals with more processing capacity to improve the computing performance and quality of service (QoS) in several typical domains of the human activity in smart society, such as social networks, medical diagnosis, telecommunications, recommendation systems, internal threat detection, transportation, Internet of Things (IoT), etc. These application domains often manage a vast collection of entities with various relationships, which can be naturally represented by the graph data structure. Graph processing is a powerful tool to model and optimize complex problems where graph-based data is involved. In consideration of the relatively insufficient resource provisioning of the edge devices, in this article, for the first time to our knowledge, we propose a reliable edge-cloud collaboration framework that facilitates the graph primitives based on a lightweight interactive graph processing library (GPL), especially for shortest path search (SPS) operations as the demonstrative example. Two types of different practical cases are also presented to show the typical application scenarios of our graph processing strategy. Experimental evaluations indicate that the acceleration rate of performance can reach 6.87x via graph reduction, and less than 3% and 20% extra latency is required for much better user experiences for navigation and pandemic control, respectively, while the online security measures merely consume about 1% extra time of the overall data transmission. Our framework can efficiently execute the applications with considering of user-friendliness, low-latency response, interactions among edge devices, collaboration between edge and cloud, and privacy protection at an acceptable overhead.","PeriodicalId":13156,"journal":{"name":"IEEE Transactions on Emerging Topics in Computing","volume":null,"pages":null},"PeriodicalIF":5.9,"publicationDate":"2023-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"62529060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Emerging Topics in Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1