首页 > 最新文献

IEEE Transactions on Sustainable Computing最新文献

英文 中文
Unsupervised Multi-Target Cross-Service Log Anomaly Detection 无监督多目标跨服务日志异常检测
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-06-10 DOI: 10.1109/TSUSC.2025.3578517
Shiming He;Rui Liu;Bowen Chen;Kun Xie;Jigang Wen
Log analysis, especially log anomaly detection, can help debug systems and analyze root causes to provide reliable services. Deep learning is a promising technology for log anomaly detection. However, deep learning methods need a large amount of training data, which is hard for a newly deployed system to collect sufficient logs. Transfer learning becomes a possible method to solve the problem that can apply the knowledge from a long-term deployed system (source) to a newly deployed system (target). Existing transfer learning methods focus on transferring the knowledge from a source system to a single target system within the same service, in which the source and the target belong to the same service (e.g. operating system, supercomputer, or distributed system). They achieve low performance when applied to multiple target and different services systems because of the obvious differences in log format, syntax, semantics, and component call between different services and the individual training of multiple models for each target system. To tackle the problems, we propose an unsupervised multi-target cross-service log anomaly detection method based on transfer learning and contrastive learning (LogMTC). LogMTC exploits contrastive learning to learn a single model on combined data from the source and multiple target systems, which can fit multiple target systems simultaneously and improve efficiency. LogMTC exploits a hypersphere loss and two contrastive losses to minimize the feature differences crossing different services. Our experiments on two services (supercomputer and distributed system) and three log datasets show that our method is superior to the existing transfer learning methods in the same service, cross-service, and multi-target log anomaly detection. Compared with the best peer accurate transfer learning algorithm LogTAD, LogMTC improves 1.14%–8.28$%$ F1 score in multi-target transfer and is 1.12–1.22 times faster.
日志分析,特别是日志异常检测,可以帮助系统调试,分析原因,提供可靠的服务。深度学习是一种很有前途的测井异常检测技术。然而,深度学习方法需要大量的训练数据,这对于新部署的系统来说很难收集到足够的日志。将长期部署的系统(源)中的知识应用到新部署的系统(目标)中,迁移学习成为解决问题的一种可能方法。现有的迁移学习方法侧重于将知识从源系统转移到同一服务内的单个目标系统,其中源系统和目标系统属于同一服务(例如操作系统、超级计算机或分布式系统)。由于不同服务之间的日志格式、语法、语义和组件调用存在明显差异,并且每个目标系统都需要对多个模型进行单独训练,因此它们在应用于多个目标和不同服务系统时性能较低。为了解决这些问题,提出了一种基于迁移学习和对比学习(LogMTC)的无监督多目标跨服务日志异常检测方法。LogMTC利用对比学习的方法,在源数据和多个目标系统的组合数据上学习单个模型,可以同时拟合多个目标系统,提高效率。LogMTC利用一个超球损耗和两个对比损耗来最小化跨不同服务的特性差异。在两种服务(超级计算机和分布式系统)和三种日志数据集上进行的实验表明,该方法在同一服务、跨服务和多目标日志异常检测方面优于现有的迁移学习方法。与最佳的同伴精确迁移学习算法LogTAD相比,LogMTC在多目标迁移中的F1分数提高了1.14% ~ 8.28%,速度提高了1.12 ~ 1.22倍。
{"title":"Unsupervised Multi-Target Cross-Service Log Anomaly Detection","authors":"Shiming He;Rui Liu;Bowen Chen;Kun Xie;Jigang Wen","doi":"10.1109/TSUSC.2025.3578517","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3578517","url":null,"abstract":"Log analysis, especially log anomaly detection, can help debug systems and analyze root causes to provide reliable services. Deep learning is a promising technology for log anomaly detection. However, deep learning methods need a large amount of training data, which is hard for a newly deployed system to collect sufficient logs. Transfer learning becomes a possible method to solve the problem that can apply the knowledge from a long-term deployed system (source) to a newly deployed system (target). Existing transfer learning methods focus on transferring the knowledge from a source system to a single target system within the same service, in which the source and the target belong to the same service (e.g. operating system, supercomputer, or distributed system). They achieve low performance when applied to multiple target and different services systems because of the obvious differences in log format, syntax, semantics, and component call between different services and the individual training of multiple models for each target system. To tackle the problems, we propose an unsupervised multi-target cross-service log anomaly detection method based on transfer learning and contrastive learning (LogMTC). LogMTC exploits contrastive learning to learn a single model on combined data from the source and multiple target systems, which can fit multiple target systems simultaneously and improve efficiency. LogMTC exploits a hypersphere loss and two contrastive losses to minimize the feature differences crossing different services. Our experiments on two services (supercomputer and distributed system) and three log datasets show that our method is superior to the existing transfer learning methods in the same service, cross-service, and multi-target log anomaly detection. Compared with the best peer accurate transfer learning algorithm LogTAD, LogMTC improves 1.14%–8.28<inline-formula><tex-math>$%$</tex-math></inline-formula> F1 score in multi-target transfer and is 1.12–1.22 times faster.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"1056-1069"},"PeriodicalIF":3.9,"publicationDate":"2025-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145248086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Efficient Scheduling Approach for Target Coverage in Solar Powered Internet of Things 太阳能物联网目标覆盖的高效调度方法
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-06-10 DOI: 10.1109/TSUSC.2025.3578433
Dipak Kumar Sah;Abhishek Hazra;Nabajyoti Mazumdar;Annavarapu Chandra Sekhara Rao;Tarachand Amgoth
The Internet of Things (IoT) has been increasingly applied in various applications in recent years. In IoT, many tasks are performed for a load operation, such as creating a cluster, preserving convergence/connectivity issues, etc. However, the energy consumption rate is also high due to more traffic in dense networks. Generally, a traditional IoT node’s battery power capacity is limited due to a short-range cycle. To address the energy shortage problem, researchers have tackled it through the Solar Powered (SP) energy harvesting technique. This method provides abundant energy to the IoT nodes at a lower cost. One issue arises from the target coverage area, which requires that each target must have at least one node for continuous monitoring in a given area. To address these issues, we have designed an effective solution called the Efficient Scheduling Target Coverage (ESTC) algorithm. This approach consists of various cover sets that work in an interleaving way. If only some node sets need to be active to satisfy coverage constraints, then there is no need to activate all sets simultaneously. ESTC provides robust coverage awareness with a perpetual network lifetime using scheduling techniques. Furthermore, the proposed work also promotes a green IoT network.
近年来,物联网(IoT)在各种应用中得到了越来越多的应用。在物联网中,许多任务都是为了负载操作而执行的,例如创建集群,保持融合/连接问题等。然而,在密集的网络中,由于更多的流量,能耗率也很高。一般来说,传统物联网节点的电池容量由于周期短而受到限制。为了解决能源短缺问题,研究人员已经通过太阳能(SP)能量收集技术解决了这个问题。该方法以较低的成本为物联网节点提供充足的能量。一个问题来自目标覆盖区域,它要求每个目标必须至少有一个节点,以便在给定区域内进行连续监视。为了解决这些问题,我们设计了一个有效的解决方案,称为高效调度目标覆盖(ESTC)算法。这种方法由以交错方式工作的各种覆盖集组成。如果只有一些节点集需要激活来满足覆盖约束,那么就没有必要同时激活所有的节点集。ESTC使用调度技术提供了具有永久网络生命周期的健壮的覆盖意识。此外,提出的工作还促进了绿色物联网网络。
{"title":"An Efficient Scheduling Approach for Target Coverage in Solar Powered Internet of Things","authors":"Dipak Kumar Sah;Abhishek Hazra;Nabajyoti Mazumdar;Annavarapu Chandra Sekhara Rao;Tarachand Amgoth","doi":"10.1109/TSUSC.2025.3578433","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3578433","url":null,"abstract":"The Internet of Things (IoT) has been increasingly applied in various applications in recent years. In IoT, many tasks are performed for a load operation, such as creating a cluster, preserving convergence/connectivity issues, etc. However, the energy consumption rate is also high due to more traffic in dense networks. Generally, a traditional IoT node’s battery power capacity is limited due to a short-range cycle. To address the energy shortage problem, researchers have tackled it through the Solar Powered (SP) energy harvesting technique. This method provides abundant energy to the IoT nodes at a lower cost. One issue arises from the target coverage area, which requires that each target must have at least one node for continuous monitoring in a given area. To address these issues, we have designed an effective solution called the Efficient Scheduling Target Coverage (ESTC) algorithm. This approach consists of various cover sets that work in an interleaving way. If only some node sets need to be active to satisfy coverage constraints, then there is no need to activate all sets simultaneously. ESTC provides robust coverage awareness with a perpetual network lifetime using scheduling techniques. Furthermore, the proposed work also promotes a green IoT network.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"1043-1055"},"PeriodicalIF":3.9,"publicationDate":"2025-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145248062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GreeNX: An Energy-Efficient and Sustainable Approach to Sparse Graph Convolution Networks Accelerators Using DVFS GreeNX:一种基于DVFS的高效、可持续的稀疏图卷积网络加速器
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-06-09 DOI: 10.1109/TSUSC.2025.3577218
Siqin Liu;Prakash Chand Kuve;Avinash Karanth
Graph convolutional networks (GCNs) have emerged as an effective approach to extend deep learning algorithms for graph-based data analytics. However, GCNs implementation over large, sparse datasets presents challenges due to irregular computation and dataflow patterns. Specialized GCN accelerators have emerged to deliver superior performance over generic processors. However, prior techniques that include specialized datapaths, optimized sparse computation, and memory access patterns, handle different phases of GCNs differently which results in excess energy consumption and reduced throughput due to sub-optimal dataflows. In this paper, we propose GreeNX, a computation and communication-aware GCN accelerator that uniformly applies three complementary techniques to all phases of GCN. First, we abstract two cascaded sparse-dense matrix multiplications that uniformly process the computation in both aggregation and combination phases of GCNs to improve throughput. Second, to mitigate the overheads of processing irregular sparse data, we develop a dynamic-voltage-and-frequency-scaling (DVFS) scheme by grouping a row of processing elements (PEs) that dynamically changes the applied voltage/frequency (V/F) to improve energy efficiency. Third, we conduct a comprehensive carbon footprint evaluation, analyzing both embodied and operational emissions for GCNs. Extensive simulation and experiments validate that our GreeNX consistently reduces memory accesses and energy consumption leading to an average 7.3× speedup and 5.6× energy savings on six real-world graph datasets over several state-of-the-art GCN accelerators including HyGCN, AWB-GCN, GCoD, GRIP, IGCN, and LW-GCN.
图卷积网络(GCNs)已成为扩展深度学习算法用于基于图的数据分析的有效方法。然而,由于不规则的计算和数据流模式,在大型稀疏数据集上实现GCNs面临挑战。专门的GCN加速器已经出现,以提供优于通用处理器的性能。然而,先前的技术包括专门的数据路径、优化的稀疏计算和内存访问模式,以不同的方式处理GCNs的不同阶段,这导致了由于次优数据流而导致的过多的能量消耗和吞吐量降低。在本文中,我们提出了GreeNX,一个计算和通信感知的GCN加速器,它统一地将三种互补技术应用于GCN的所有阶段。首先,我们抽象了两个级联的稀疏密集矩阵乘法,在GCNs的聚合和组合阶段统一处理计算,以提高吞吐量。其次,为了减轻处理不规则稀疏数据的开销,我们开发了一种动态电压和频率缩放(DVFS)方案,该方案通过分组一排动态改变施加电压/频率(V/F)的处理元素(pe)来提高能源效率。第三,我们进行了全面的碳足迹评估,分析了GCNs的隐含和运营排放。大量的仿真和实验验证了我们的GreeNX持续减少内存访问和能耗,在几种最先进的GCN加速器(包括HyGCN, AWB-GCN, GCoD, GRIP, IGCN和LW-GCN)上,在六个真实世界的图形数据集上平均加速7.3倍,节能5.6倍。
{"title":"GreeNX: An Energy-Efficient and Sustainable Approach to Sparse Graph Convolution Networks Accelerators Using DVFS","authors":"Siqin Liu;Prakash Chand Kuve;Avinash Karanth","doi":"10.1109/TSUSC.2025.3577218","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3577218","url":null,"abstract":"Graph convolutional networks (GCNs) have emerged as an effective approach to extend deep learning algorithms for graph-based data analytics. However, GCNs implementation over large, sparse datasets presents challenges due to irregular computation and dataflow patterns. Specialized GCN accelerators have emerged to deliver superior performance over generic processors. However, prior techniques that include specialized datapaths, optimized sparse computation, and memory access patterns, handle different phases of GCNs differently which results in excess energy consumption and reduced throughput due to sub-optimal dataflows. In this paper, we propose GreeNX, a computation and communication-aware GCN accelerator that uniformly applies three complementary techniques to all phases of GCN. First, we abstract two cascaded sparse-dense matrix multiplications that uniformly process the computation in both aggregation and combination phases of GCNs to improve throughput. Second, to mitigate the overheads of processing irregular sparse data, we develop a dynamic-voltage-and-frequency-scaling (DVFS) scheme by grouping a row of processing elements (PEs) that dynamically changes the applied voltage/frequency (V/F) to improve energy efficiency. Third, we conduct a comprehensive carbon footprint evaluation, analyzing both embodied and operational emissions for GCNs. Extensive simulation and experiments validate that our GreeNX consistently reduces memory accesses and energy consumption leading to an average 7.3× speedup and 5.6× energy savings on six real-world graph datasets over several state-of-the-art GCN accelerators including HyGCN, AWB-GCN, GCoD, GRIP, IGCN, and LW-GCN.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"1031-1042"},"PeriodicalIF":3.9,"publicationDate":"2025-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145248071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HS-GCN: A High-Performance, Sustainable, and Scalable Chiplet-Based Accelerator for Graph Convolutional Network Inference HS-GCN:一种高性能、可持续、可扩展的基于芯片的图卷积网络推理加速器
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-06-02 DOI: 10.1109/TSUSC.2025.3575285
Yingnan Zhao;Ke Wang;Ahmed Louri
Graph Convolutional Networks (GCNs) have been proposed to extend machine learning techniques for graph-related applications. A typical GCN model consists of multiple layers, each including an aggregation phase, which is communication-intensive, and a combination phase, which is computation-intensive. As the size of real-world graphs increases exponentially, current customized accelerators face challenges in efficiently performing GCN inference due to limited on-chip buffers and other hardware resources for both data computation and communication, which degrades performance and energy efficiency. Additionally, scaling current monolithic designs to address the aforementioned challenges will introduce significant cost-effectiveness issues in terms of power, area, and yield. To this end, we propose HS-GCN, a high-performance, sustainable, and scalable chiplet-based accelerator for GCN inference with much-improved energy efficiency. Specifically, HS-GCN integrates multiple reconfigurable chiplets, each of which can be configured to perform the main computations of either the aggregation phase or the combination phase, including Sparse-dense matrix multiplication (SpMM) and General matrix-matrix multiplication (GeMM). HS-GCN implements an active interposer with a flexible interconnection fabric to connect chiplets and other hardware components for efficient data communication. Additionally, HS-GCN introduces two system-level control algorithms that dynamically determine the computation order and corresponding dataflow based on the input graphs and GCN models. These selections are used to further configure the chiplet array and interconnection fabric for much-improved performance and energy efficiency. Evaluation results using real-world graphs demonstrate that HS-GCN achieves significant speedups of 26.7×, 11.2×, 3.9×, 4.7×, 3.1×, along with substantial memory access savings of 94%, 89%, 64%, 85%, 54%, and energy savings of 87%, 84%, 49%, 78%, 41% on average, as compared to HyGCN, AWB-GCN, GCNAX, I-GCN, and SGCN, respectively.
图卷积网络(GCNs)被提出用于扩展与图相关的应用的机器学习技术。典型的GCN模型由多层组成,每层都包括一个通信密集型的聚合阶段和一个计算密集型的组合阶段。随着现实世界图形的大小呈指数级增长,由于芯片上的缓冲区和其他用于数据计算和通信的硬件资源有限,当前的定制加速器在有效执行GCN推理方面面临挑战,这会降低性能和能源效率。此外,扩展当前的单片设计以解决上述挑战将在功率,面积和良率方面引入重大的成本效益问题。为此,我们提出了HS-GCN,这是一种高性能,可持续和可扩展的基于芯片的GCN推理加速器,具有大大提高的能源效率。具体来说,HS-GCN集成了多个可重构小芯片,每个小芯片都可以配置为执行聚合阶段或组合阶段的主要计算,包括稀疏密集矩阵乘法(SpMM)和通用矩阵矩阵乘法(GeMM)。HS-GCN实现了一个具有灵活互连结构的有源中介器,用于连接小芯片和其他硬件组件,以实现高效的数据通信。此外,HS-GCN引入了两种系统级控制算法,根据输入图和GCN模型动态确定计算顺序和相应的数据流。这些选择用于进一步配置芯片阵列和互连结构,以大大提高性能和能源效率。使用真实图形的评估结果表明,与HyGCN、AWB-GCN、GCNAX、I-GCN和SGCN相比,HS-GCN实现了26.7×、11.2×、3.9×、4.7×、3.1×的显著速度提升,内存访问节省了94%、89%、64%、85%、54%,平均节能了87%、84%、49%、78%、41%。
{"title":"HS-GCN: A High-Performance, Sustainable, and Scalable Chiplet-Based Accelerator for Graph Convolutional Network Inference","authors":"Yingnan Zhao;Ke Wang;Ahmed Louri","doi":"10.1109/TSUSC.2025.3575285","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3575285","url":null,"abstract":"Graph Convolutional Networks (GCNs) have been proposed to extend machine learning techniques for graph-related applications. A typical GCN model consists of multiple layers, each including an aggregation phase, which is communication-intensive, and a combination phase, which is computation-intensive. As the size of real-world graphs increases exponentially, current customized accelerators face challenges in efficiently performing GCN inference due to limited on-chip buffers and other hardware resources for both data computation and communication, which degrades performance and energy efficiency. Additionally, scaling current monolithic designs to address the aforementioned challenges will introduce significant cost-effectiveness issues in terms of power, area, and yield. To this end, we propose HS-GCN, a high-performance, sustainable, and scalable chiplet-based accelerator for GCN inference with much-improved energy efficiency. Specifically, HS-GCN integrates multiple reconfigurable chiplets, each of which can be configured to perform the main computations of either the aggregation phase or the combination phase, including Sparse-dense matrix multiplication (SpMM) and General matrix-matrix multiplication (GeMM). HS-GCN implements an active interposer with a flexible interconnection fabric to connect chiplets and other hardware components for efficient data communication. Additionally, HS-GCN introduces two system-level control algorithms that dynamically determine the computation order and corresponding dataflow based on the input graphs and GCN models. These selections are used to further configure the chiplet array and interconnection fabric for much-improved performance and energy efficiency. Evaluation results using real-world graphs demonstrate that HS-GCN achieves significant speedups of 26.7×, 11.2×, 3.9×, 4.7×, 3.1×, along with substantial memory access savings of 94%, 89%, 64%, 85%, 54%, and energy savings of 87%, 84%, 49%, 78%, 41% on average, as compared to HyGCN, AWB-GCN, GCNAX, I-GCN, and SGCN, respectively.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"1019-1030"},"PeriodicalIF":3.9,"publicationDate":"2025-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145256014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
APPARENT: AI-Powered Platform Anomaly Detection in Edge Computing 显而易见:人工智能驱动的边缘计算平台异常检测
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-04-21 DOI: 10.1109/TSUSC.2025.3562738
Chandrajit Pal;Sangeet Saha;Xiaojun Zhai;Gareth Howells;Klaus D. McDonald-Maier
Embedded systems serving as IoT nodes are often vulnerable to malicious and unknown runtime software that could compromise the system, steal sensitive data, and cause undesirable system behaviour. Commercially available embedded systems used in automation, medical equipment, and automotive industries, are especially exposed to this vulnerability since they lack the resources to incorporate conventional safety features and are challenging to mitigate through conventional approaches. We propose a novel system design coined as APPARENT which can identify program characteristics by monitoring and counting the maximum possible low-level hardware events from Hardware Performance Counters (HPCs) that occur during the program's execution and analyse the correlation among the counts of various monitored events. To further utilise these captured events as features we propose a self-supervised machine learning algorithm that combines a Graph Attention Network GAT and a Generative Topographic Mapping GTM to detect unusual program behaviour as anomalies to enhance the system security. Our proposed methodology takes advantage of attributes like program counter, cycles per instruction, and physical and virtual timers at various exception levels of the embedded processor to identify abnormal activity. APPARENT identifies unknown program behaviours not present in the training phase with an accuracy of over 98.46% on Autobench EEMBC benchmarks.
作为物联网节点的嵌入式系统通常容易受到恶意和未知运行时软件的攻击,这些软件可能会危害系统,窃取敏感数据并导致不良的系统行为。用于自动化、医疗设备和汽车行业的商用嵌入式系统尤其容易受到此漏洞的影响,因为它们缺乏整合传统安全功能的资源,并且很难通过传统方法加以缓解。我们提出了一种新的系统设计,它可以通过监控和计数硬件性能计数器(hpc)在程序执行期间发生的最大可能的低级硬件事件来识别程序特征,并分析各种监控事件计数之间的相关性。为了进一步利用这些捕获的事件作为特征,我们提出了一种自监督机器学习算法,该算法结合了图注意网络GAT和生成式地形映射GTM来检测异常程序行为作为异常以增强系统安全性。我们提出的方法利用了诸如程序计数器、每指令周期、嵌入式处理器各种异常级别上的物理和虚拟计时器等属性来识别异常活动。在Autobench EEMBC基准测试中,obvious识别训练阶段不存在的未知程序行为,准确率超过98.46%。
{"title":"APPARENT: AI-Powered Platform Anomaly Detection in Edge Computing","authors":"Chandrajit Pal;Sangeet Saha;Xiaojun Zhai;Gareth Howells;Klaus D. McDonald-Maier","doi":"10.1109/TSUSC.2025.3562738","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3562738","url":null,"abstract":"Embedded systems serving as IoT nodes are often vulnerable to malicious and unknown runtime software that could compromise the system, steal sensitive data, and cause undesirable system behaviour. Commercially available embedded systems used in automation, medical equipment, and automotive industries, are especially exposed to this vulnerability since they lack the resources to incorporate conventional safety features and are challenging to mitigate through conventional approaches. We propose a novel system design coined as APPARENT which can identify program characteristics by monitoring and counting the maximum possible low-level hardware events from Hardware Performance Counters (HPCs) that occur during the program's execution and analyse the correlation among the counts of various monitored events. To further utilise these captured events as features we propose a self-supervised machine learning algorithm that combines a Graph Attention Network GAT and a Generative Topographic Mapping GTM to detect unusual program behaviour as anomalies to enhance the system security. Our proposed methodology takes advantage of attributes like program counter, cycles per instruction, and physical and virtual timers at various exception levels of the embedded processor to identify abnormal activity. APPARENT identifies unknown program behaviours not present in the training phase with an accuracy of over 98.46% on Autobench EEMBC benchmarks.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"965-981"},"PeriodicalIF":3.9,"publicationDate":"2025-04-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145248060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainable AI-Guided Neural Architecture Search for Adversarial Robustness in Approximate DNNs 近似dnn中对抗鲁棒性的可解释ai引导神经结构搜索
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-04-16 DOI: 10.1109/TSUSC.2025.3561603
Ayesha Siddique;Khaza Anuarul Hoque
Deep neural networks are lucrative targets of adversarial attacks and approximate deep neural networks (AxDNNs) are no exception. Searching manually for adversarially robust AxDNN architectures incurs outrageous time and human effort. In this paper, we propose XAI-NAS, an explainable neural architecture search (NAS) method that leverages explainable artificial intelligence (XAI) to efficiently co-optimize the adversarial robustness and hardware efficiency of AxDNN architectures on systolic-array hardware accelerators. During the NAS process, AxDNN architectures are evolved layer-wise with heterogeneous approximate multipliers to deliver the best trade-offs between adversarial robustness, energy consumption, latency, and memory footprint. The most suitable approximate multipliers are automatically selected from an open-source Evoapprox8b library. Our extensive evaluations provide a set of Pareto optimal hardware efficient and adversarially robust solutions. For example, a Pareto-optimal DNN AxDNN for the MNIST and CIFAR-10 datasets exhibits up to 1.5× higher adversarial robustness, 2.1× less energy consumption, 4.39× reduced latency, and 2.37× low memory footprint when compared to the state-of-the-art NAS approaches.
深度神经网络是对抗性攻击的有利目标,近似深度神经网络(axdnn)也不例外。手动搜索对抗健壮的AxDNN体系结构需要大量的时间和人力。在本文中,我们提出了XAI-NAS,一种可解释神经架构搜索(NAS)方法,利用可解释人工智能(XAI)有效地协同优化收缩阵列硬件加速器上AxDNN架构的对抗鲁棒性和硬件效率。在NAS过程中,AxDNN架构使用异构近似乘数器分层演进,以提供对抗性鲁棒性、能耗、延迟和内存占用之间的最佳权衡。最合适的近似乘数会自动从开源的Evoapprox8b库中选择。我们广泛的评估提供了一套帕累托最优硬件效率和对抗稳健的解决方案。例如,与最先进的NAS方法相比,用于MNIST和CIFAR-10数据集的pareto最优DNN AxDNN显示出高达1.5倍的对抗鲁棒性,2.1倍的能耗,4.39倍的延迟减少和2.37倍的内存占用。
{"title":"Explainable AI-Guided Neural Architecture Search for Adversarial Robustness in Approximate DNNs","authors":"Ayesha Siddique;Khaza Anuarul Hoque","doi":"10.1109/TSUSC.2025.3561603","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3561603","url":null,"abstract":"Deep neural networks are lucrative targets of adversarial attacks and approximate deep neural networks (AxDNNs) are no exception. Searching manually for adversarially robust AxDNN architectures incurs outrageous time and human effort. In this paper, we propose XAI-NAS, an explainable neural architecture search (NAS) method that leverages explainable artificial intelligence (XAI) to efficiently co-optimize the adversarial robustness and hardware efficiency of AxDNN architectures on systolic-array hardware accelerators. During the NAS process, AxDNN architectures are evolved layer-wise with heterogeneous approximate multipliers to deliver the best trade-offs between adversarial robustness, energy consumption, latency, and memory footprint. The most suitable approximate multipliers are automatically selected from an open-source Evoapprox8b library. Our extensive evaluations provide a set of Pareto optimal hardware efficient and adversarially robust solutions. For example, a Pareto-optimal DNN AxDNN for the MNIST and CIFAR-10 datasets exhibits up to 1.5× higher adversarial robustness, 2.1× less energy consumption, 4.39× reduced latency, and 2.37× low memory footprint when compared to the state-of-the-art NAS approaches.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"949-964"},"PeriodicalIF":3.9,"publicationDate":"2025-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145248094","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Novel Stealth Communication Round Attack and Robust Incentivized Federated Averaging for Load Forecasting 新型隐身通信轮攻击与鲁棒激励联邦平均负荷预测
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-14 DOI: 10.1109/TSUSC.2025.3570096
Habib Ullah Manzoor;Kamran Arshad;Khaled Assaleh;Ahmed Zoha
Federated learning (FL) has gained prominence in energy forecasting applications. Despite its advantages, FL remains vulnerable to adversarial attacks that threaten the reliability of predictive models. This study introduces a stealth attack, Federated Communication Round Attack (Fed-CRA), which increases communication rounds without affecting forecasting accuracy. Increased communication rounds can delay decision-making, reducing system responsiveness and cost-effectiveness in dynamic energy forecasting scenarios. Experimental validation on two datasets demonstrated that Fed-CRA increased communication rounds by 574% (from 72 to 485) in the AEP dataset and by 237% (from 92 to 310) in the COMED dataset. This led to a corresponding rise in energy consumption by 573% (from 41.04 kWh to 276.35 kWh) and 237% (from 52.44 kWh to 176.65 kWh), respectively, while preserving forecasting accuracy. To counter this attack, we proposed Federated Incentivized Averaging (Fed-InA), a game theory-inspired framework that rewards honest clients and penalizes dishonest ones based on their contributions. Results showed that Fed-InA reduced the additional communication rounds caused by Fed-CRA by 85% in the AEP dataset and 70% in the COMED dataset, while maintaining forecasting performance. Fed-InA achieves resource efficiency comparable to Federated Averaging (FedAvg) and demonstrates robustness in handling non-IID data.
联邦学习(FL)在能源预测应用中得到了突出的应用。尽管具有优势,但FL仍然容易受到威胁预测模型可靠性的对抗性攻击。本研究引入了一种隐形攻击,联邦通信回合攻击(federal Communication Round attack, Fed-CRA),它在不影响预测精度的情况下增加了通信回合。在动态能源预测场景中,增加的通信轮次可能会延迟决策,降低系统响应能力和成本效益。两个数据集的实验验证表明,Fed-CRA在AEP数据集中增加了574%(从72次增加到485次),在COMED数据集中增加了237%(从92次增加到310次)。在保持预测准确性的前提下,能耗相应增加573%(从41.04 kWh增加到276.35 kWh)和237%(从52.44 kWh增加到176.65 kWh)。为了应对这种攻击,我们提出了联邦激励平均(Fed-InA),这是一个博弈论启发的框架,奖励诚实的客户,并根据他们的贡献惩罚不诚实的客户。结果表明,在保持预测性能的同时,Fed-InA在AEP数据集中减少了85%由Fed-CRA引起的额外沟通轮数,在COMED数据集中减少了70%。Fed-InA实现了与联邦平均(fedag)相当的资源效率,并在处理非iid数据方面表现出鲁棒性。
{"title":"Novel Stealth Communication Round Attack and Robust Incentivized Federated Averaging for Load Forecasting","authors":"Habib Ullah Manzoor;Kamran Arshad;Khaled Assaleh;Ahmed Zoha","doi":"10.1109/TSUSC.2025.3570096","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3570096","url":null,"abstract":"Federated learning (FL) has gained prominence in energy forecasting applications. Despite its advantages, FL remains vulnerable to adversarial attacks that threaten the reliability of predictive models. This study introduces a stealth attack, Federated Communication Round Attack (Fed-CRA), which increases communication rounds without affecting forecasting accuracy. Increased communication rounds can delay decision-making, reducing system responsiveness and cost-effectiveness in dynamic energy forecasting scenarios. Experimental validation on two datasets demonstrated that Fed-CRA increased communication rounds by 574% (from 72 to 485) in the AEP dataset and by 237% (from 92 to 310) in the COMED dataset. This led to a corresponding rise in energy consumption by 573% (from 41.04 kWh to 276.35 kWh) and 237% (from 52.44 kWh to 176.65 kWh), respectively, while preserving forecasting accuracy. To counter this attack, we proposed Federated Incentivized Averaging (Fed-InA), a game theory-inspired framework that rewards honest clients and penalizes dishonest ones based on their contributions. Results showed that Fed-InA reduced the additional communication rounds caused by Fed-CRA by 85% in the AEP dataset and 70% in the COMED dataset, while maintaining forecasting performance. Fed-InA achieves resource efficiency comparable to Federated Averaging (FedAvg) and demonstrates robustness in handling non-IID data.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"1007-1018"},"PeriodicalIF":3.9,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145248080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Wireless Rechargeable Sensor Networks: Energy Provisioning Technologies, Charging Scheduling Schemes, and Challenges 无线可充电传感器网络:能源供应技术、充电调度方案和挑战
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-10 DOI: 10.1109/TSUSC.2025.3549414
Samah Abdel Aziz;Xingfu Wang;Ammar Hawbani;Bushra Qureshi;Saeed H. Alsamhi;Aisha Alabsi;Liang Zhao;Ahmed Al-Dubai;A.S. Ismail
Recently, a plethora of promising green energy provisioning technologies has been discussed in the orientation of prolonging the lifetime of energy-limited devices (e.g., sensor nodes). Wireless rechargeable sensor networks (WRSNs) have emerged among other fields that could greatly benefit from such technologies. Such an ad-hoc network comprises a base station(s) and multiple sensor nodes, which are primarily deployed in harsh environments, meeting the requirements of transmitting, receiving, collecting, and processing data. Unlike existing works, this survey paper focuses on energy provisioning technologies within the context of WRSNs by reviewing two interrelated domains. First, we introduce various energy provisioning techniques and their associated challenges, including conventional energy harvesting methods (e.g., solar, thermal, and mechanical). We highlight wireless power transfer (WPT) as one of the most applicable technologies for WRSNs, covering both radiative and non-radiative WPT. Additionally, we present radio frequency (RF) energy harvesting, including simultaneous wireless information and power transfer (SWIPT) and wireless powered communication networks (WPCNs), as well as backscatter communications. Furthermore, we compare hybrid energy harvesting techniques (e.g., solar-RF, vibro-acoustic, solar-thermal, etc.). Second, we introduce the fundamentals of wireless charging, reviewing various charger types (static and mobile), charging policies (including full and partial charging), charging modes (offline and online), and charging schemes (periodic and on-demand). We also present the collaborative charging mechanisms. Additionally, we address several key challenges facing WRSNs, such as energy consumption, multi-charger coordination, dynamic network recharging, monitoring & security threats, vehicle-to-vehicle (V2V) charging, and hybrid WRSNs Finally, we highlight trends and future directions for integrating advanced artificial intelligence (AI) technologies into WRSNs.
最近,在延长能量受限设备(如传感器节点)寿命的方向上,讨论了大量有前途的绿色能源供应技术。无线可充电传感器网络(WRSNs)已经出现在其他领域,可以从这种技术中受益匪浅。这种自组织网络由一个基站和多个传感器节点组成,主要部署在恶劣环境中,满足数据的发送、接收、采集和处理要求。与现有工作不同,本调查论文通过回顾两个相互关联的领域,重点关注wrns背景下的能源供应技术。首先,我们介绍了各种能量供应技术及其相关的挑战,包括传统的能量收集方法(例如,太阳能,热能和机械)。无线功率传输(WPT)是wrns中最适用的技术之一,包括辐射和非辐射WPT。此外,我们还介绍了射频(RF)能量收集,包括同步无线信息和电力传输(SWIPT)和无线供电通信网络(wpcn),以及反向散射通信。此外,我们比较了混合能量收集技术(例如,太阳能射频,振动声,太阳能热等)。其次,我们介绍了无线充电的基本原理,回顾了各种充电器类型(静态和移动),充电策略(包括完全充电和部分充电),充电模式(离线和在线)以及充电方案(定期和按需)。我们还提出了协同收费机制。此外,我们还讨论了WRSNs面临的几个关键挑战,如能源消耗、多充电器协调、动态网络充电、监控和安全威胁、车对车(V2V)充电和混合WRSNs。最后,我们强调了将先进人工智能(AI)技术集成到WRSNs的趋势和未来方向。
{"title":"Wireless Rechargeable Sensor Networks: Energy Provisioning Technologies, Charging Scheduling Schemes, and Challenges","authors":"Samah Abdel Aziz;Xingfu Wang;Ammar Hawbani;Bushra Qureshi;Saeed H. Alsamhi;Aisha Alabsi;Liang Zhao;Ahmed Al-Dubai;A.S. Ismail","doi":"10.1109/TSUSC.2025.3549414","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3549414","url":null,"abstract":"Recently, a plethora of promising green energy provisioning technologies has been discussed in the orientation of prolonging the lifetime of energy-limited devices (e.g., sensor nodes). Wireless rechargeable sensor networks (WRSNs) have emerged among other fields that could greatly benefit from such technologies. Such an ad-hoc network comprises a base station(s) and multiple sensor nodes, which are primarily deployed in harsh environments, meeting the requirements of transmitting, receiving, collecting, and processing data. Unlike existing works, this survey paper focuses on energy provisioning technologies within the context of WRSNs by reviewing two interrelated domains. First, we introduce various energy provisioning techniques and their associated challenges, including conventional energy harvesting methods (e.g., solar, thermal, and mechanical). We highlight wireless power transfer (WPT) as one of the most applicable technologies for WRSNs, covering both radiative and non-radiative WPT. Additionally, we present radio frequency (RF) energy harvesting, including simultaneous wireless information and power transfer (SWIPT) and wireless powered communication networks (WPCNs), as well as backscatter communications. Furthermore, we compare hybrid energy harvesting techniques (e.g., solar-RF, vibro-acoustic, solar-thermal, etc.). Second, we introduce the fundamentals of wireless charging, reviewing various charger types (static and mobile), charging policies (including full and partial charging), charging modes (offline and online), and charging schemes (periodic and on-demand). We also present the collaborative charging mechanisms. Additionally, we address several key challenges facing WRSNs, such as energy consumption, multi-charger coordination, dynamic network recharging, monitoring & security threats, vehicle-to-vehicle (V2V) charging, and hybrid WRSNs Finally, we highlight trends and future directions for integrating advanced artificial intelligence (AI) technologies into WRSNs.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"873-890"},"PeriodicalIF":3.9,"publicationDate":"2025-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145248058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
$x$xPUE: Extending Power Usage Effectiveness Metrics For Cloud Infrastructures $x$xPUE:扩展云基础设施的电力使用效率指标
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-10 DOI: 10.1109/TSUSC.2025.3549687
Guillaume Fieni;Romain Rouvoy;Lionel Seinturier
The energy consumption analysis and optimization of data centers have been an increasingly popular topic over the past few years. It is widely recognized that several effective metrics exist to capture the efficiency of hardware and/or software hosted in these infrastructures. Unfortunately, choosing the corresponding metrics for specific infrastructure and assessing its efficiency over time is still considered an open problem. For this purpose, energy efficiency metrics, such as the Power Usage Effectiveness (PUE), assess the efficiency of the computing equipment of the infrastructure. However, this metric stops at the power supply of hosted servers and fails to offer a finer granularity to bring a deeper insight into the Power Usage Effectiveness of hardware and software running in cloud infrastructure. Therefore, we propose to leverage complementary PUE metrics, coined $x$PUE, to compute the energy efficiency of the computing continuum from hardware components, up to the running software layers. Our contribution aims to deliver real-time energy efficiency metrics from different perspectives for cloud infrastructure, hence helping cloud ecosystems—from cloud providers to their customers—to experiment and optimize the energy usage of cloud infrastructures at large.
在过去的几年里,数据中心的能耗分析和优化已经成为一个越来越流行的话题。人们普遍认识到,存在一些有效的度量来捕获这些基础设施中托管的硬件和/或软件的效率。不幸的是,为特定的基础设施选择相应的指标并评估其随时间的效率仍然被认为是一个悬而未决的问题。为此,能源效率指标,如电力使用效率(PUE),评估基础设施的计算设备的效率。然而,该指标停留在托管服务器的电源上,无法提供更细的粒度来更深入地了解云基础设施中运行的硬件和软件的电源使用效率。因此,我们建议利用互补的PUE指标,即$x$PUE,来计算从硬件组件到运行软件层的计算连续体的能源效率。我们的贡献旨在从不同的角度为云基础设施提供实时能源效率指标,从而帮助云生态系统——从云提供商到他们的客户——试验和优化云基础设施的能源使用。
{"title":"$x$xPUE: Extending Power Usage Effectiveness Metrics For Cloud Infrastructures","authors":"Guillaume Fieni;Romain Rouvoy;Lionel Seinturier","doi":"10.1109/TSUSC.2025.3549687","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3549687","url":null,"abstract":"The energy consumption analysis and optimization of data centers have been an increasingly popular topic over the past few years. It is widely recognized that several effective metrics exist to capture the efficiency of hardware and/or software hosted in these infrastructures. Unfortunately, choosing the corresponding metrics for specific infrastructure and assessing its efficiency over time is still considered an open problem. For this purpose, energy efficiency metrics, such as the <i>Power Usage Effectiveness</i> (PUE), assess the efficiency of the computing equipment of the infrastructure. However, this metric stops at the power supply of hosted servers and fails to offer a finer granularity to bring a deeper insight into the <i>Power Usage Effectiveness</i> of hardware and software running in cloud infrastructure. Therefore, we propose to leverage complementary PUE metrics, coined <inline-formula><tex-math>$x$</tex-math></inline-formula>PUE, to compute the energy efficiency of the computing continuum from hardware components, up to the running software layers. Our contribution aims to deliver real-time energy efficiency metrics from different perspectives for cloud infrastructure, hence helping cloud ecosystems—from cloud providers to their customers—to experiment and optimize the energy usage of cloud infrastructures at large.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"908-920"},"PeriodicalIF":3.9,"publicationDate":"2025-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145248043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data-Driven Software-Based Power Estimation for Embedded Devices 基于数据驱动软件的嵌入式设备功率估计
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-08 DOI: 10.1109/TSUSC.2025.3567856
Haoyu Wang;Xinyi Li;Ti Zhou;Man Lin
Energy measurement of computer devices, which are widely used in the Internet of Things (IoT), is an important yet challenging task. Most of these IoT devices lack ready-to-use hardware or software for power measurement. In this paper, we propose an easy-to-use approach to derive a software-based energy estimation model with external low-end power meters based on data-driven analysis. Our solution is demonstrated with a Jetson Nano board and Ruideng UM25C USB power meter. Various machine learning methods combined with our smart data collection & profiling method and physical measurement are explored. Periodic Long-duration measurements are utilized in the experiments to derive and validate power models, allowing more accurate power readings from the low-end power meter. Benchmarks were used to evaluate the derived software-power model for the Jetson Nano board and Raspberry Pi. The results show that 92% accuracy can be achieved by the software-based power estimation compared to measurement. A kernel module that can collect running traces of utilization and frequencies needed is developed, together with the power model derived, for power prediction for programs running in a real environment. Our cost-effective method facilitates accurate instantaneous power estimation, which low-end power meters cannot directly provide.
在物联网(IoT)中广泛应用的计算机设备的能量测量是一项重要而富有挑战性的任务。这些物联网设备大多缺乏现成的硬件或软件来测量功率。在本文中,我们提出了一种易于使用的方法,基于数据驱动分析,推导出基于软件的外部低端功率表的能量估算模型。我们的解决方案用Jetson Nano板和瑞登UM25C USB功率计进行了演示。探索各种机器学习方法与我们的智能数据收集和分析方法以及物理测量相结合。在实验中使用周期性长时间测量来推导和验证功率模型,从而从低端功率计获得更准确的功率读数。基准测试用于评估Jetson Nano板和树莓派的衍生软件功率模型。结果表明,与实测相比,基于软件的功率估计精度可达92%。开发了一个内核模块,可以收集程序的运行轨迹和所需的频率,并建立了功耗模型,用于实际环境中运行的程序的功耗预测。我们的经济高效的方法可以实现准确的瞬时功率估计,这是低端电能表无法直接提供的。
{"title":"Data-Driven Software-Based Power Estimation for Embedded Devices","authors":"Haoyu Wang;Xinyi Li;Ti Zhou;Man Lin","doi":"10.1109/TSUSC.2025.3567856","DOIUrl":"https://doi.org/10.1109/TSUSC.2025.3567856","url":null,"abstract":"Energy measurement of computer devices, which are widely used in the Internet of Things (IoT), is an important yet challenging task. Most of these IoT devices lack ready-to-use hardware or software for power measurement. In this paper, we propose an easy-to-use approach to derive a software-based energy estimation model with external low-end power meters based on data-driven analysis. Our solution is demonstrated with a Jetson Nano board and Ruideng UM25C USB power meter. Various machine learning methods combined with our smart data collection & profiling method and physical measurement are explored. Periodic Long-duration measurements are utilized in the experiments to derive and validate power models, allowing more accurate power readings from the low-end power meter. Benchmarks were used to evaluate the derived software-power model for the Jetson Nano board and Raspberry Pi. The results show that 92% accuracy can be achieved by the software-based power estimation compared to measurement. A kernel module that can collect running traces of utilization and frequencies needed is developed, together with the power model derived, for power prediction for programs running in a real environment. Our cost-effective method facilitates accurate instantaneous power estimation, which low-end power meters cannot directly provide.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 5","pages":"937-948"},"PeriodicalIF":3.9,"publicationDate":"2025-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145248099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Sustainable Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1