首页 > 最新文献

BenchCouncil Transactions on Benchmarks, Standards and Evaluations最新文献

英文 中文
ERMDS: A obfuscation dataset for evaluating robustness of learning-based malware detection system 一个用于评估基于学习的恶意软件检测系统鲁棒性的混淆数据集
Pub Date : 2023-02-01 DOI: 10.1016/j.tbench.2023.100106
Lichen Jia , Yang Yang , Bowen Tang , Zihan Jiang

Learning-based malware detection systems (LB-MDS) play a crucial role in defending computer systems from malicious attacks. Nevertheless, these systems can be vulnerable to various attacks, which can have significant consequences. Software obfuscation techniques can be used to modify the features of malware, thereby avoiding its classification as malicious by LB-MDS. However, existing portable executable (PE) malware datasets primarily use a single obfuscation technique, which LB-MDS has already learned, leading to a loss of their robustness evaluation ability. Therefore, creating a dataset with diverse features that were not observed during LB-MDS training has become the main challenge in evaluating the robustness of LB-MDS.

We propose a obfuscation dataset ERMDS that solves the problem of evaluating the robustness of LB-MDS by generating malwares with diverse features. When designing this dataset, we created three types of obfuscation spaces, corresponding to binary obfuscation, source code obfuscation, and packing obfuscation. Each obfuscation space has multiple obfuscation techniques, each with different parameters. The obfuscation techniques in these three obfuscation spaces can be used in combination and can be reused. This enables us to theoretically obtain an infinite number of obfuscation combinations, thereby creating malwares with a diverse range of features that have not been captured by LB-MDS.

To assess the effectiveness of the ERMDS obfuscation dataset, we create an instance of the obfuscation dataset called ERMDS-X. By utilizing this dataset, we conducted an evaluation of the robustness of two LB-MDS models, namely MalConv and EMBER, as well as six commercial antivirus software products, which are anonymized as AV1-AV6. The results of our experiments showed that ERMDS-X effectively reveals the limitations in the robustness of existing LB-MDS models, leading to an average accuracy reduction of 20% in LB-MDS and 32% in commercial antivirus software. We conducted a comprehensive analysis of the factors that contributed to the observed accuracy decline in both LB-MDS and commercial antivirus software. We have released the ERMDS-X dataset as an open-source resource, available on GitHub at https://github.com/lcjia94/ERMDS.

基于学习的恶意软件检测系统(LB-MDS)在保护计算机系统免受恶意攻击方面发挥着至关重要的作用。然而,这些系统可能容易受到各种攻击,从而产生重大后果。软件混淆技术可用于修改恶意软件的特征,从而避免其被LB-MDS归类为恶意软件。然而,现有的可移植可执行(PE)恶意软件数据集主要使用LB-MDS已经学会的单一模糊技术,导致其鲁棒性评估能力的丧失。因此,创建一个具有在LB-MDS训练过程中没有观察到的各种特征的数据集已成为评估LB-MDS鲁棒性的主要挑战。我们提出了一个模糊数据集ERMDS,该数据集通过生成具有各种特征的恶意软件来解决评估LB-MDS-鲁棒性的问题。在设计该数据集时,我们创建了三种类型的模糊空间,分别对应于二进制模糊、源代码模糊和打包模糊。每个模糊处理空间都有多种模糊处理技术,每种技术都有不同的参数。这三个模糊空间中的模糊技术可以组合使用,并且可以重用。这使我们能够在理论上获得无限数量的模糊组合,从而创建具有LB-MDS尚未捕获的各种功能的恶意软件。为了评估ERMDS模糊数据集的有效性,我们创建了一个名为ERMDS-X的模糊数据集实例。通过利用该数据集,我们对两个LB-MDS模型(即MalConv和EMBR)以及六个商业杀毒软件产品(匿名为AV1-AV6)的稳健性进行了评估。我们的实验结果表明,ERMDS-X有效地揭示了现有LB-MDS模型稳健性的局限性,导致LB-MDS的平均准确率降低了20%,商业反病毒软件的平均准确度降低了32%。我们对LB-MDS和商业杀毒软件中导致观察到的准确性下降的因素进行了全面分析。我们已经发布了ERMDS-X数据集作为开源资源,可在GitHub上获得,网址为https://github.com/lcjia94/ERMDS.
{"title":"ERMDS: A obfuscation dataset for evaluating robustness of learning-based malware detection system","authors":"Lichen Jia ,&nbsp;Yang Yang ,&nbsp;Bowen Tang ,&nbsp;Zihan Jiang","doi":"10.1016/j.tbench.2023.100106","DOIUrl":"https://doi.org/10.1016/j.tbench.2023.100106","url":null,"abstract":"<div><p>Learning-based malware detection systems (LB-MDS) play a crucial role in defending computer systems from malicious attacks. Nevertheless, these systems can be vulnerable to various attacks, which can have significant consequences. Software obfuscation techniques can be used to modify the features of malware, thereby avoiding its classification as malicious by LB-MDS. However, existing portable executable (PE) malware datasets primarily use a single obfuscation technique, which LB-MDS has already learned, leading to a loss of their robustness evaluation ability. Therefore, creating a dataset with diverse features that were not observed during LB-MDS training has become the main challenge in evaluating the robustness of LB-MDS.</p><p>We propose a obfuscation dataset ERMDS that solves the problem of evaluating the robustness of LB-MDS by generating malwares with diverse features. When designing this dataset, we created three types of obfuscation spaces, corresponding to binary obfuscation, source code obfuscation, and packing obfuscation. Each obfuscation space has multiple obfuscation techniques, each with different parameters. The obfuscation techniques in these three obfuscation spaces can be used in combination and can be reused. This enables us to theoretically obtain an infinite number of obfuscation combinations, thereby creating malwares with a diverse range of features that have not been captured by LB-MDS.</p><p>To assess the effectiveness of the ERMDS obfuscation dataset, we create an instance of the obfuscation dataset called ERMDS-X. By utilizing this dataset, we conducted an evaluation of the robustness of two LB-MDS models, namely MalConv and EMBER, as well as six commercial antivirus software products, which are anonymized as AV1-AV6. The results of our experiments showed that ERMDS-X effectively reveals the limitations in the robustness of existing LB-MDS models, leading to an average accuracy reduction of 20% in LB-MDS and 32% in commercial antivirus software. We conducted a comprehensive analysis of the factors that contributed to the observed accuracy decline in both LB-MDS and commercial antivirus software. We have released the ERMDS-X dataset as an open-source resource, available on GitHub at <span>https://github.com/lcjia94/ERMDS</span><svg><path></path></svg>.</p></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"3 1","pages":"Article 100106"},"PeriodicalIF":0.0,"publicationDate":"2023-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49714566","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enabling hyperscale web services 启用超大规模web服务
Pub Date : 2023-02-01 DOI: 10.1016/j.tbench.2023.100092
Akshitha Sriraman

Modern web services such as social media, online messaging, and web search support billions of users, requiring data centers that scale to hundreds of thousands of servers, i.e., hyperscale. The key challenge in enabling hyperscale web services arise from (1) an unprecedented growth in data, users, and service functionality and (2) a decline in hardware performance scaling. We highlight a dissertation’s contributions in bridging the software and hardware worlds to realize more efficient hyperscale services despite these challenges.

社交媒体、在线消息和网络搜索等现代网络服务支持数十亿用户,需要扩展到数十万服务器的数据中心,即超大规模。实现超大规模web服务的关键挑战来自(1)数据、用户和服务功能的空前增长,以及(2)硬件性能扩展的下降。我们强调了一篇论文在连接软件和硬件世界以实现更高效的超规模服务方面的贡献,尽管存在这些挑战。
{"title":"Enabling hyperscale web services","authors":"Akshitha Sriraman","doi":"10.1016/j.tbench.2023.100092","DOIUrl":"https://doi.org/10.1016/j.tbench.2023.100092","url":null,"abstract":"<div><p>Modern web services such as social media, online messaging, and web search support billions of users, requiring data centers that scale to hundreds of thousands of servers, i.e., <em>hyperscale</em>. The key challenge in enabling hyperscale web services arise from (1) an unprecedented growth in data, users, and service functionality and (2) a decline in hardware performance scaling. We highlight a dissertation’s contributions in bridging the software and hardware worlds to realize more efficient hyperscale services despite these challenges.</p></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"3 1","pages":"Article 100092"},"PeriodicalIF":0.0,"publicationDate":"2023-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49714575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing the sparse approximate inverse preconditioning algorithm on GPU 基于GPU的稀疏近似逆预处理算法优化
Pub Date : 2022-10-01 DOI: 10.1016/j.tbench.2023.100087
Xinyue Chu, Yizhou Wang, Qi Chen, Jiaquan Gao

In this study, we present an optimization sparse approximate inverse (SPAI) preconditioning algorithm on GPU, called GSPAI-Opt. In GSPAI-Opt, it fuses the advantages of two popular SPAI preconditioning algorithms, and has the following novelties: (1) an optimization strategy is proposed to choose whether to use the constant or non-constant thread group for any sparse pattern of the preprocessor, and (2) a parallel framework of optimizing the SPAI preconditioner is proposed on GPU, and (3) for each component of the preconditioner, a decision tree is established to choose the optimal kernel of computing it. Experimental results validate the effectiveness of GSPAI-Opt.

在本研究中,我们提出了一种GPU上的优化稀疏近似逆(SPAI)预处理算法,称为GSPAI-Opt。GSPAI-Opt,融合两种流行的优势SPAI预处理算法,和下面的小礼品:(1)提出了一种优化策略选择是否使用常数或不恒定线程组的稀疏模式预处理器,和(2)一个并行的框架上,提出了优化SPAI预处理GPU,和(3)对于每个组件的预处理,建立决策树来选择最优计算内核。实验结果验证了GSPAI-Opt算法的有效性。
{"title":"Optimizing the sparse approximate inverse preconditioning algorithm on GPU","authors":"Xinyue Chu,&nbsp;Yizhou Wang,&nbsp;Qi Chen,&nbsp;Jiaquan Gao","doi":"10.1016/j.tbench.2023.100087","DOIUrl":"10.1016/j.tbench.2023.100087","url":null,"abstract":"<div><p>In this study, we present an optimization sparse approximate inverse (SPAI) preconditioning algorithm on GPU, called GSPAI-Opt. In GSPAI-Opt, it fuses the advantages of two popular SPAI preconditioning algorithms, and has the following novelties: (1) an optimization strategy is proposed to choose whether to use the constant or non-constant thread group for any sparse pattern of the preprocessor, and (2) a parallel framework of optimizing the SPAI preconditioner is proposed on GPU, and (3) for each component of the preconditioner, a decision tree is established to choose the optimal kernel of computing it. Experimental results validate the effectiveness of GSPAI-Opt.</p></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"2 4","pages":"Article 100087"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2772485923000042/pdfft?md5=8592fe298c854dc9f2e85112414f0c44&pid=1-s2.0-S2772485923000042-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77940299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Performance characterization and optimization of pruning patterns for sparse DNN inference 稀疏DNN推理中剪枝模式的性能表征与优化
Pub Date : 2022-10-01 DOI: 10.1016/j.tbench.2023.100090
Yunjie Liu, Jingwei Sun, Jiaqiang Liu, Guangzhong Sun

Deep neural networks are suffering from over parameterized high storage and high consumption problems. Pruning can effectively reduce storage and computation costs of deep neural networks by eliminating their redundant parameters. In existing pruning methods, filter pruning achieves more efficient inference, while element-wise pruning maintains better accuracy. To make a trade-off between the two endpoints, a variety of pruning patterns has been proposed. This study analyzes the performance characteristics of sparse DNNs pruned by different patterns, including element-wise, vector-wise, block-wise, and group-wise. Based on the analysis, we propose an efficient implementation of group-wise sparse DNN inference, which can make better use of GPUs. Experimental results on VGG, ResNet, BERT and ViT show that our optimized group-wise pruning pattern achieves much lower inference latency on GPU than other sparse patterns and the existing group-wise pattern implementation.

深度神经网络存在过参数化的高存储和高消耗问题。剪枝通过消除冗余参数,有效地降低了深度神经网络的存储和计算成本。在现有的剪枝方法中,过滤器剪枝可以实现更高效的推理,而元素剪枝可以保持更好的准确性。为了在两个端点之间进行权衡,提出了各种修剪模式。本研究分析了不同模式下的稀疏dnn的性能特征,包括元素型、矢量型、块型和组型。在此基础上,我们提出了一种有效的分组稀疏DNN推理实现方法,可以更好地利用gpu。在VGG、ResNet、BERT和ViT上的实验结果表明,优化后的组明智修剪模式在GPU上的推理延迟比其他稀疏模式和现有的组明智模式实现要低得多。
{"title":"Performance characterization and optimization of pruning patterns for sparse DNN inference","authors":"Yunjie Liu,&nbsp;Jingwei Sun,&nbsp;Jiaqiang Liu,&nbsp;Guangzhong Sun","doi":"10.1016/j.tbench.2023.100090","DOIUrl":"10.1016/j.tbench.2023.100090","url":null,"abstract":"<div><p>Deep neural networks are suffering from over parameterized high storage and high consumption problems. Pruning can effectively reduce storage and computation costs of deep neural networks by eliminating their redundant parameters. In existing pruning methods, filter pruning achieves more efficient inference, while element-wise pruning maintains better accuracy. To make a trade-off between the two endpoints, a variety of pruning patterns has been proposed. This study analyzes the performance characteristics of sparse DNNs pruned by different patterns, including element-wise, vector-wise, block-wise, and group-wise. Based on the analysis, we propose an efficient implementation of group-wise sparse DNN inference, which can make better use of GPUs. Experimental results on VGG, ResNet, BERT and ViT show that our optimized group-wise pruning pattern achieves much lower inference latency on GPU than other sparse patterns and the existing group-wise pattern implementation.</p></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"2 4","pages":"Article 100090"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2772485923000078/pdfft?md5=47f436d7570515bb39cfffeda4376c89&pid=1-s2.0-S2772485923000078-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84120145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TBench (BenchCouncil Transactions on Benchmarks, Standards and Evaluations) Calls for Papers bench (BenchCouncil Transactions on benchmark, Standards and evaluation)征文
Pub Date : 2022-10-01 DOI: 10.1016/j.tbench.2023.100103
{"title":"TBench (BenchCouncil Transactions on Benchmarks, Standards and Evaluations) Calls for Papers","authors":"","doi":"10.1016/j.tbench.2023.100103","DOIUrl":"10.1016/j.tbench.2023.100103","url":null,"abstract":"","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"2 4","pages":"Article 100103"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2772485923000200/pdfft?md5=ee27feb6477e417eae72f364987d955c&pid=1-s2.0-S2772485923000200-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90528369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IoTBench: A data centrical and configurable IoT benchmark suite IoTBench:以数据为中心和可配置的物联网基准套件
Pub Date : 2022-10-01 DOI: 10.1016/j.tbench.2023.100091
Simin Chen , Chunjie Luo , Wanling Gao , Lei Wang

As the Internet of Things (IoT) industry expands, the demand for microprocessors and microcontrollers used in IoT systems has increased steadily. Benchmarks provide a valuable reference for processor evaluation. Different IoT application scenarios face different data scales, dimensions, and types. However, the current popular benchmarks only evaluate the processor’s performance under fixed data formats. These benchmarks cannot adapt to the fragmented scenarios faced by processors. This paper proposes a new benchmark, namely IoTBench. The IoTBench workloads cover three types of algorithms commonly used in IoT applications: matrix processing, list operation, and convolution. Moreover, IoTBench divides the data space into different evaluation subspaces according to the data scales, data types, and data dimensions. We analyze the impact of different data types, data dimensions, and data scales on processor performance and compare ARM with RISC-V and MinorCPU with O3CPU using IoTBench. We also explored the performance of processors with different architecture configurations in different evaluation subspaces and found the optimal architecture of different evaluation subspaces. The specifications, source code, and results are publicly available from https://www.benchcouncil.org/iotbench/.

随着物联网(IoT)行业的扩张,对物联网系统中使用的微处理器和微控制器的需求稳步增长。基准测试为处理器评估提供了有价值的参考。不同的物联网应用场景面临不同的数据规模、维度和类型。然而,目前流行的基准测试只评估处理器在固定数据格式下的性能。这些基准测试不能适应处理器所面临的分散场景。本文提出了一种新的基准,即IoTBench。IoTBench工作负载涵盖物联网应用中常用的三种算法:矩阵处理、列表操作和卷积。此外,IoTBench根据数据规模、数据类型和数据维度将数据空间划分为不同的评估子空间。我们分析了不同数据类型、数据维度和数据规模对处理器性能的影响,并使用IoTBench比较了ARM与RISC-V和MinorCPU与O3CPU。探讨了不同架构配置的处理器在不同求值子空间中的性能,找到了不同求值子空间的最优架构。规范、源代码和结果可从https://www.benchcouncil.org/iotbench/公开获得。
{"title":"IoTBench: A data centrical and configurable IoT benchmark suite","authors":"Simin Chen ,&nbsp;Chunjie Luo ,&nbsp;Wanling Gao ,&nbsp;Lei Wang","doi":"10.1016/j.tbench.2023.100091","DOIUrl":"10.1016/j.tbench.2023.100091","url":null,"abstract":"<div><p>As the Internet of Things (IoT) industry expands, the demand for microprocessors and microcontrollers used in IoT systems has increased steadily. Benchmarks provide a valuable reference for processor evaluation. Different IoT application scenarios face different data scales, dimensions, and types. However, the current popular benchmarks only evaluate the processor’s performance under fixed data formats. These benchmarks cannot adapt to the fragmented scenarios faced by processors. This paper proposes a new benchmark, namely IoTBench. The IoTBench workloads cover three types of algorithms commonly used in IoT applications: matrix processing, list operation, and convolution. Moreover, IoTBench divides the data space into different evaluation subspaces according to the data scales, data types, and data dimensions. We analyze the impact of different data types, data dimensions, and data scales on processor performance and compare ARM with RISC-V and MinorCPU with O3CPU using IoTBench. We also explored the performance of processors with different architecture configurations in different evaluation subspaces and found the optimal architecture of different evaluation subspaces. The specifications, source code, and results are publicly available from <span>https://www.benchcouncil.org/iotbench/</span><svg><path></path></svg>.</p></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"2 4","pages":"Article 100091"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S277248592300008X/pdfft?md5=3e608f0131eab9659bc377156487a717&pid=1-s2.0-S277248592300008X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76076182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A review of Blockchain Technology applications for financial services 区块链技术在金融服务中的应用综述
Pub Date : 2022-10-01 DOI: 10.1016/j.tbench.2022.100073
M. Javaid, Abid Haleem, R. Singh, R. Suman, Shahbaz Khan
{"title":"A review of Blockchain Technology applications for financial services","authors":"M. Javaid, Abid Haleem, R. Singh, R. Suman, Shahbaz Khan","doi":"10.1016/j.tbench.2022.100073","DOIUrl":"https://doi.org/10.1016/j.tbench.2022.100073","url":null,"abstract":"","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"407 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75474486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Edge AIBench 2.0: A scalable autonomous vehicle benchmark for IoT–Edge–Cloud systems Edge AIBench 2.0:物联网边缘云系统的可扩展自动驾驶汽车基准
Pub Date : 2022-10-01 DOI: 10.1016/j.tbench.2023.100086
Tianshu Hao , Wanling Gao , Chuanxin Lan , Fei Tang , Zihan Jiang , Jianfeng Zhan

Many emerging IoT–Edge–Cloud computing systems are not yet implemented or are too confidential to share the code or even tricky to replicate its execution environment, and hence their benchmarking is very challenging. This paper uses autonomous vehicles as a typical scenario to build the first benchmark for IoT–Edge–Cloud systems. We propose a set of distilling rules for replicating autonomous vehicle scenarios to extract critical tasks with intertwined interactions. The essential system-level and component-level characteristics are captured while the system complexity is reduced significantly so that users can quickly evaluate and pinpoint the system and component bottlenecks. Also, we implement a scalable architecture through which users can assess the systems with different sizes of workloads.

We conduct several experiments to measure the performance. After testing two thousand autonomous vehicle task requests, we identify the bottleneck modules in autonomous vehicle scenarios and analyze their hotspot functions. The experiment results show that the lane-keeping task is the slowest execution module, with a tail latency of 77.49 ms for the 99th percentile latency. We hope this scenario benchmark will be helpful for Autonomous Vehicles and even IoT–edge–Cloud research. Now the open-source code is available from the official website https://www.benchcouncil.org/scenariobench/edgeaibench.html.

许多新兴的物联网边缘云计算系统尚未实现,或者过于机密而无法共享代码,甚至难以复制其执行环境,因此它们的基准测试非常具有挑战性。本文以自动驾驶汽车为典型场景,构建物联网边缘云系统的第一个基准。我们提出了一套用于复制自动驾驶汽车场景的提取规则,以提取相互交织的关键任务。在显著降低系统复杂性的同时,捕获了基本的系统级和组件级特征,以便用户可以快速评估和查明系统和组件瓶颈。此外,我们还实现了一个可扩展的体系结构,用户可以通过该体系结构评估具有不同工作负载大小的系统。我们进行了几个实验来衡量性能。在测试了2000个自动驾驶汽车任务请求后,我们确定了自动驾驶汽车场景中的瓶颈模块,并分析了它们的热点功能。实验结果表明,车道保持任务是执行速度最慢的模块,尾部延迟为77.49 ms,为第99百分位延迟。我们希望这个场景基准将对自动驾驶汽车甚至物联网边缘云研究有所帮助。现在可以从官方网站https://www.benchcouncil.org/scenariobench/edgeaibench.html获得开源代码。
{"title":"Edge AIBench 2.0: A scalable autonomous vehicle benchmark for IoT–Edge–Cloud systems","authors":"Tianshu Hao ,&nbsp;Wanling Gao ,&nbsp;Chuanxin Lan ,&nbsp;Fei Tang ,&nbsp;Zihan Jiang ,&nbsp;Jianfeng Zhan","doi":"10.1016/j.tbench.2023.100086","DOIUrl":"10.1016/j.tbench.2023.100086","url":null,"abstract":"<div><p>Many emerging IoT–Edge–Cloud computing systems are not yet implemented or are too confidential to share the code or even tricky to replicate its execution environment, and hence their benchmarking is very challenging. This paper uses autonomous vehicles as a typical scenario to build the first benchmark for IoT–Edge–Cloud systems. We propose a set of distilling rules for replicating autonomous vehicle scenarios to extract critical tasks with intertwined interactions. The essential system-level and component-level characteristics are captured while the system complexity is reduced significantly so that users can quickly evaluate and pinpoint the system and component bottlenecks. Also, we implement a scalable architecture through which users can assess the systems with different sizes of workloads.</p><p>We conduct several experiments to measure the performance. After testing two thousand autonomous vehicle task requests, we identify the bottleneck modules in autonomous vehicle scenarios and analyze their hotspot functions. The experiment results show that the lane-keeping task is the slowest execution module, with a tail latency of 77.49 ms for the 99th percentile latency. We hope this scenario benchmark will be helpful for Autonomous Vehicles and even IoT–edge–Cloud research. Now the open-source code is available from the official website <span>https://www.benchcouncil.org/scenariobench/edgeaibench.html</span><svg><path></path></svg>.</p></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"2 4","pages":"Article 100086"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2772485923000030/pdfft?md5=f59a880c243b7557a7fcd0ca689dd1e8&pid=1-s2.0-S2772485923000030-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76164703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Diagnosis of COVID-19 from X-rays using combined CNN-RNN architecture with transfer learning 结合CNN-RNN架构和迁移学习的x射线COVID-19诊断
Pub Date : 2022-10-01 DOI: 10.1016/j.tbench.2023.100088
Md. Milon Islam , Md. Zabirul Islam , Amanullah Asraf , Mabrook S. Al-Rakhami , Weiping Ding , Ali Hassan Sodhro

Combating the COVID-19 pandemic has emerged as one of the most promising issues in global healthcare. Accurate and fast diagnosis of COVID-19 cases is required for the right medical treatment to control this pandemic. Chest radiography imaging techniques are more effective than the reverse-transcription polymerase chain reaction (RT-PCR) method in detecting coronavirus. Due to the limited availability of medical images, transfer learning is better suited to classify patterns in medical images. This paper presents a combined architecture of convolutional neural network (CNN) and recurrent neural network (RNN) to diagnose COVID-19 patients from chest X-rays. The deep transfer techniques used in this experiment are VGG19, DenseNet121, InceptionV3, and Inception-ResNetV2, where CNN is used to extract complex features from samples and classify them using RNN. In our experiments, the VGG19-RNN architecture outperformed all other networks in terms of accuracy. Finally, decision-making regions of images were visualized using gradient-weighted class activation mapping (Grad-CAM). The system achieved promising results compared to other existing systems and might be validated in the future when more samples would be available. The experiment demonstrated a good alternative method to diagnose COVID-19 for medical staff.

All the data used during the study are openly available from the Mendeley data repository at https://data.mendeley.com/datasets/mxc6vb7svm. For further research, we have made the source code publicly available at https://github.com/Asraf047/COVID19-CNN-RNN.

抗击COVID-19大流行已成为全球医疗保健领域最有希望的问题之一。需要准确和快速诊断COVID-19病例,以便采取正确的医疗措施来控制这场大流行。胸片成像技术在检测冠状病毒方面比逆转录聚合酶链反应(RT-PCR)方法更有效。由于医学图像的可用性有限,迁移学习更适合于医学图像中的模式分类。本文提出了一种卷积神经网络(CNN)和递归神经网络(RNN)的组合架构,用于从胸部x光片诊断COVID-19患者。本实验中使用的深度转移技术有VGG19、DenseNet121、inception - resnetv3和Inception-ResNetV2,其中使用CNN从样本中提取复杂特征,并使用RNN进行分类。在我们的实验中,VGG19-RNN架构在准确率方面优于所有其他网络。最后,利用梯度加权类激活映射(Grad-CAM)对图像的决策区域进行可视化。与其他现有系统相比,该系统取得了令人鼓舞的结果,并可能在未来更多样品可用时进行验证。该实验为医务人员提供了一种很好的替代诊断方法。研究过程中使用的所有数据都可以从Mendeley数据库中公开获取,网址为https://data.mendeley.com/datasets/mxc6vb7svm。为了进一步研究,我们在https://github.com/Asraf047/COVID19-CNN-RNN上公开了源代码。
{"title":"Diagnosis of COVID-19 from X-rays using combined CNN-RNN architecture with transfer learning","authors":"Md. Milon Islam ,&nbsp;Md. Zabirul Islam ,&nbsp;Amanullah Asraf ,&nbsp;Mabrook S. Al-Rakhami ,&nbsp;Weiping Ding ,&nbsp;Ali Hassan Sodhro","doi":"10.1016/j.tbench.2023.100088","DOIUrl":"10.1016/j.tbench.2023.100088","url":null,"abstract":"<div><p>Combating the COVID-19 pandemic has emerged as one of the most promising issues in global healthcare. Accurate and fast diagnosis of COVID-19 cases is required for the right medical treatment to control this pandemic. Chest radiography imaging techniques are more effective than the reverse-transcription polymerase chain reaction (RT-PCR) method in detecting coronavirus. Due to the limited availability of medical images, transfer learning is better suited to classify patterns in medical images. This paper presents a combined architecture of convolutional neural network (CNN) and recurrent neural network (RNN) to diagnose COVID-19 patients from chest X-rays. The deep transfer techniques used in this experiment are VGG19, DenseNet121, InceptionV3, and Inception-ResNetV2, where CNN is used to extract complex features from samples and classify them using RNN. In our experiments, the VGG19-RNN architecture outperformed all other networks in terms of accuracy. Finally, decision-making regions of images were visualized using gradient-weighted class activation mapping (Grad-CAM). The system achieved promising results compared to other existing systems and might be validated in the future when more samples would be available. The experiment demonstrated a good alternative method to diagnose COVID-19 for medical staff.</p><p>All the data used during the study are openly available from the Mendeley data repository at <span>https://data.mendeley.com/datasets/mxc6vb7svm</span><svg><path></path></svg>. For further research, we have made the source code publicly available at <span>https://github.com/Asraf047/COVID19-CNN-RNN</span><svg><path></path></svg>.</p></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"2 4","pages":"Article 100088"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2772485923000054/pdfft?md5=b7e11074cc12a010ea7f32743ca0e2a5&pid=1-s2.0-S2772485923000054-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81905664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 73
Enabling Reduced Simpoint Size Through LiveCache and Detail Warmup 通过LiveCache和细节预热来减小Simpoint大小
Pub Date : 2022-10-01 DOI: 10.1016/j.tbench.2022.100082
Jose Renau , Fangping Liu , Hongzhang Shan , Sang Wook Stephen Do

Simpoint technology (Sherwood et al., 2002) has been widely used by modern micro-architecture research community to significantly speedup the simulation time. However, the typical Simpoint size remains to be tens to hundreds of million instructions. At such sizes, the cycle-accurate simulators still need to run tens of hours or even days to finish the simulation, depending on the architecture complexity and workload characteristics. In this paper, we developed a new simulation framework by integrating LiveCache and Detail-warmups with Dromajo ( https://chipyard.readthedocs.io/en/latest/Tools/Dromajo.html) and Kabylkas et al. (2005), enabling us to use much smaller Simpoint size (2 million instructions) without loss of accuracy. Our evaluation results showed that the average simulation time can be accelerated by 9.56 times over 50M size and most of the workload simulations can be finished in tens of minutes instead of hours.

Simpoint技术(Sherwood et al., 2002)被现代微架构研究界广泛使用,可以显著加快仿真时间。然而,典型的Simpoint大小仍然是数千万到数亿条指令。在这样的规模下,周期精确的模拟器仍然需要运行数十小时甚至几天才能完成模拟,这取决于体系结构复杂性和工作负载特征。在本文中,我们通过将LiveCache和细节预热与Dromajo (https://chipyard.readthedocs.io/en/latest/Tools/Dromajo.html)和Kabylkas等人(2005)集成开发了一个新的模拟框架,使我们能够使用更小的Simpoint大小(200万条指令)而不会损失准确性。我们的评估结果表明,在50M大小的情况下,平均模拟时间可以加快9.56倍,大多数工作负载的模拟可以在几十分钟内完成,而不是几个小时。
{"title":"Enabling Reduced Simpoint Size Through LiveCache and Detail Warmup","authors":"Jose Renau ,&nbsp;Fangping Liu ,&nbsp;Hongzhang Shan ,&nbsp;Sang Wook Stephen Do","doi":"10.1016/j.tbench.2022.100082","DOIUrl":"10.1016/j.tbench.2022.100082","url":null,"abstract":"<div><p>Simpoint technology (Sherwood et al., 2002) has been widely used by modern micro-architecture research community to significantly speedup the simulation time. However, the typical Simpoint size remains to be tens to hundreds of million instructions. At such sizes, the cycle-accurate simulators still need to run tens of hours or even days to finish the simulation, depending on the architecture complexity and workload characteristics. In this paper, we developed a new simulation framework by integrating LiveCache and Detail-warmups with Dromajo ( <span>https://chipyard.readthedocs.io/en/latest/Tools/Dromajo.html</span><svg><path></path></svg>) and Kabylkas et al. (2005), enabling us to use much smaller Simpoint size (2 million instructions) without loss of accuracy. Our evaluation results showed that the average simulation time can be accelerated by 9.56 times over 50M size and most of the workload simulations can be finished in tens of minutes instead of hours.</p></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"2 4","pages":"Article 100082"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2772485922000692/pdfft?md5=1f63e574d9398dea849e86896a519786&pid=1-s2.0-S2772485922000692-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83497373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
BenchCouncil Transactions on Benchmarks, Standards and Evaluations
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1