首页 > 最新文献

Proceedings of the 2021 on Systems and Network Telemetry and Analytics最新文献

英文 中文
Session details: Keynote 2 会议详情:主题演讲2
Pub Date : 2020-06-21 DOI: 10.1145/3240508.3286918
Jinoh Kim
{"title":"Session details: Keynote 2","authors":"Jinoh Kim","doi":"10.1145/3240508.3286918","DOIUrl":"https://doi.org/10.1145/3240508.3286918","url":null,"abstract":"","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127486626","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Characterizing Resource Heterogeneity in Edge Devices for Deep Learning Inferences 基于深度学习推理的边缘设备资源异构特征
Pub Date : 2020-06-21 DOI: 10.1145/3452411.3464446
Jianwei Hao, Piyush Subedi, I. Kim, Lakshmish Ramaswamy
Significant advances in hardware capabilities and the availability of enormous data sets have led to the rise and penetration of artificial intelligence (AI) and deep learning (DL) in various domains. Considerable efforts have been put forth in academia and industry to make these computationally demanding DL tasks work on resource-constrained edge devices. However, performing DL tasks on edge devices is still challenging due to the diversity of DNN (Deep Neural Networks) architectures and heterogeneity of edge devices. This study evaluates and characterizes the performance and resource heterogeneity in various edge devices for performing DL tasks. We benchmark various DNN models for image classification on a set of edge devices ranging from the widely popular and relatively less powerful Raspberry Pi to GPU-equipped high-performance edge devices like Jetson Xavier NX. We also compare and contrast the performance of three widely-used DL frameworks when used in these edge devices. We report DL inference throughput, CPU and memory usage, power consumption, and frameworks' initialization overhead, which are the most critical factors for characterizing DL tasks on edge devices. Additionally, we provide our insights and findings, which will provide a better idea of how compatible or feasible edge devices are for running DL applications.
硬件能力的显著进步和海量数据集的可用性导致了人工智能(AI)和深度学习(DL)在各个领域的兴起和渗透。学术界和工业界已经付出了相当大的努力,使这些计算要求很高的深度学习任务能够在资源受限的边缘设备上工作。然而,由于DNN(深度神经网络)架构的多样性和边缘设备的异质性,在边缘设备上执行深度学习任务仍然具有挑战性。本研究评估和表征了执行深度学习任务的各种边缘设备的性能和资源异质性。我们在一组边缘设备上对各种DNN模型进行了图像分类基准测试,这些设备从广受欢迎但功能相对较弱的树莓派到配备gpu的高性能边缘设备,如Jetson Xavier NX。我们还比较和对比了在这些边缘设备中使用的三种广泛使用的深度学习框架的性能。我们报告了深度学习推理吞吐量、CPU和内存使用、功耗和框架初始化开销,这些是表征边缘设备上深度学习任务的最关键因素。此外,我们还提供了我们的见解和发现,这将更好地了解边缘设备对于运行DL应用程序的兼容性或可行性。
{"title":"Characterizing Resource Heterogeneity in Edge Devices for Deep Learning Inferences","authors":"Jianwei Hao, Piyush Subedi, I. Kim, Lakshmish Ramaswamy","doi":"10.1145/3452411.3464446","DOIUrl":"https://doi.org/10.1145/3452411.3464446","url":null,"abstract":"Significant advances in hardware capabilities and the availability of enormous data sets have led to the rise and penetration of artificial intelligence (AI) and deep learning (DL) in various domains. Considerable efforts have been put forth in academia and industry to make these computationally demanding DL tasks work on resource-constrained edge devices. However, performing DL tasks on edge devices is still challenging due to the diversity of DNN (Deep Neural Networks) architectures and heterogeneity of edge devices. This study evaluates and characterizes the performance and resource heterogeneity in various edge devices for performing DL tasks. We benchmark various DNN models for image classification on a set of edge devices ranging from the widely popular and relatively less powerful Raspberry Pi to GPU-equipped high-performance edge devices like Jetson Xavier NX. We also compare and contrast the performance of three widely-used DL frameworks when used in these edge devices. We report DL inference throughput, CPU and memory usage, power consumption, and frameworks' initialization overhead, which are the most critical factors for characterizing DL tasks on edge devices. Additionally, we provide our insights and findings, which will provide a better idea of how compatible or feasible edge devices are for running DL applications.","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127086203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Session details: Technical Session 1 会议详情:技术会议1
M. Cafaro
{"title":"Session details: Technical Session 1","authors":"M. Cafaro","doi":"10.1145/3470766","DOIUrl":"https://doi.org/10.1145/3470766","url":null,"abstract":"","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127648060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
INODE - Intelligence Open Data Exploration 智能开放数据探索
Pub Date : 2020-06-21 DOI: 10.1145/3452411.3464448
Kurt Stockinger
This article describes the keynote speech on INODE presented at Fourth International Workshop on Systems and Network Telemetry and Analytics (SNTA) which is collocated with International ACM Symposium on High-Performance Parallel and Distributed Computing (HPDC) on June 21 in Stockholm, Sweden.
本文描述了6月21日在瑞典斯德哥尔摩举行的第四届系统和网络遥测与分析国际研讨会(SNTA)上关于INODE的主题演讲,该研讨会与国际ACM高性能并行和分布式计算研讨会(HPDC)同时举行。
{"title":"INODE - Intelligence Open Data Exploration","authors":"Kurt Stockinger","doi":"10.1145/3452411.3464448","DOIUrl":"https://doi.org/10.1145/3452411.3464448","url":null,"abstract":"This article describes the keynote speech on INODE presented at Fourth International Workshop on Systems and Network Telemetry and Analytics (SNTA) which is collocated with International ACM Symposium on High-Performance Parallel and Distributed Computing (HPDC) on June 21 in Stockholm, Sweden.","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116721448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hybrid Virtual Network Function Placement Strategy for Maximizing the Profit of Network Service Deployment over Dynamic Workload 动态负载下网络服务部署利润最大化的混合虚拟网络功能布局策略
Pub Date : 2020-06-21 DOI: 10.1145/3452411.3464440
Chi-Chen Yang, J. Chou
The emergence of network function virtualization~(NFV) has revolutionized the infrastructure and service management of network architecture by reducing the cost and complexity of network service deployment. However, finding the optimal placement of virtual network functions (VNFs) is an NP-complete problem. Existing solutions base on either Integer Linear Programming~(ILP), or greedy algorithms. But, solving ILP can be time consuming and the approximation of greedy is not bounded. Hence, neither of them can make quick and accurate placement decisions, especially under dynamic traffic workload. Therefore, we propose a hybrid method that combines the two approaches together to achieve up to 45% service profit improvement with less computation time comparing to the traditional static ILP approach.
网络功能虚拟化(NFV)的出现通过降低网络服务部署的成本和复杂性,彻底改变了网络架构的基础设施和服务管理。然而,寻找虚拟网络函数(VNFs)的最优位置是一个np完全问题。现有的解要么基于整数线性规划~(ILP),要么基于贪婪算法。但是,求解ILP是非常耗时的,而且贪心的近似是无界的。因此,它们都不能快速准确地做出布局决策,特别是在动态交通负载下。因此,我们提出了一种将两种方法结合在一起的混合方法,与传统的静态ILP方法相比,计算时间更少,服务利润提高高达45%。
{"title":"A Hybrid Virtual Network Function Placement Strategy for Maximizing the Profit of Network Service Deployment over Dynamic Workload","authors":"Chi-Chen Yang, J. Chou","doi":"10.1145/3452411.3464440","DOIUrl":"https://doi.org/10.1145/3452411.3464440","url":null,"abstract":"The emergence of network function virtualization~(NFV) has revolutionized the infrastructure and service management of network architecture by reducing the cost and complexity of network service deployment. However, finding the optimal placement of virtual network functions (VNFs) is an NP-complete problem. Existing solutions base on either Integer Linear Programming~(ILP), or greedy algorithms. But, solving ILP can be time consuming and the approximation of greedy is not bounded. Hence, neither of them can make quick and accurate placement decisions, especially under dynamic traffic workload. Therefore, we propose a hybrid method that combines the two approaches together to achieve up to 45% service profit improvement with less computation time comparing to the traditional static ILP approach.","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133403266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analyzing Scientific Data Sharing Patterns for In-network Data Caching 网络内数据缓存科学数据共享模式分析
Pub Date : 2020-06-21 DOI: 10.1145/3452411.3464441
Elizabeth Copps, Huiyi Zhang, A. Sim, Kesheng Wu, I. Monga, C. Guok, F. Würthwein, Diego Davila, E. Hernandez
The volume of data moving through a network increases with new scientific experiments and simulations. Network bandwidth requirements also increase proportionally to deliver data within a certain time frame. We observe that a significant portion of the popular dataset is transferred multiple times to different users as well as to the same user for various reasons. In-network data caching for the shared data has shown to reduce the redundant data transfers and consequently save network traffic volume. In addition, overall application performance is expected to improve with in-network caching because access to the locally cached data results in lower latency. This paper shows how much data was shared over the study period, how much network traffic volume was consequently saved, and how much the temporary in-network caching increased the scientific application performance. It also analyzes data access patterns in applications and the impacts of caching nodes on the regional data repository. From the results, we observed that the network bandwidth demand was reduced by nearly a factor of 3 over the study period.
随着新的科学实验和模拟,通过网络传输的数据量也在增加。为了在特定的时间范围内传输数据,网络带宽需求也会成比例地增加。我们观察到,由于各种原因,流行数据集的很大一部分被多次传输给不同的用户以及同一用户。共享数据的网络内数据缓存已被证明可以减少冗余数据传输,从而节省网络流量。此外,使用网络内缓存可以提高应用程序的整体性能,因为访问本地缓存的数据可以降低延迟。本文展示了在研究期间共享了多少数据,因此节省了多少网络流量,以及临时网络内缓存提高了多少科学应用程序性能。本文还分析了应用程序中的数据访问模式以及缓存节点对区域数据存储库的影响。从结果中,我们观察到,在研究期间,网络带宽需求减少了近3倍。
{"title":"Analyzing Scientific Data Sharing Patterns for In-network Data Caching","authors":"Elizabeth Copps, Huiyi Zhang, A. Sim, Kesheng Wu, I. Monga, C. Guok, F. Würthwein, Diego Davila, E. Hernandez","doi":"10.1145/3452411.3464441","DOIUrl":"https://doi.org/10.1145/3452411.3464441","url":null,"abstract":"The volume of data moving through a network increases with new scientific experiments and simulations. Network bandwidth requirements also increase proportionally to deliver data within a certain time frame. We observe that a significant portion of the popular dataset is transferred multiple times to different users as well as to the same user for various reasons. In-network data caching for the shared data has shown to reduce the redundant data transfers and consequently save network traffic volume. In addition, overall application performance is expected to improve with in-network caching because access to the locally cached data results in lower latency. This paper shows how much data was shared over the study period, how much network traffic volume was consequently saved, and how much the temporary in-network caching increased the scientific application performance. It also analyzes data access patterns in applications and the impacts of caching nodes on the regional data repository. From the results, we observed that the network bandwidth demand was reduced by nearly a factor of 3 over the study period.","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121438033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Session details: Technical Session 2 会议详情:技术会议2
Jinoh Kim
{"title":"Session details: Technical Session 2","authors":"Jinoh Kim","doi":"10.1145/3470768","DOIUrl":"https://doi.org/10.1145/3470768","url":null,"abstract":"","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"2020 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117058631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Recent Advances and Future Challenges for Network Function Virtualization Infrastructure 网络功能虚拟化基础设施的最新进展和未来挑战
Pub Date : 2020-06-21 DOI: 10.1145/3452411.3464449
J. Chou
Today's enterprise networks has revolutionized by the emerging technology called Network function virtualization (NFV), which is a type of data center network architecture proposed by the European Telecommunications Standards Institute (ETSI). NFV uses virtualization techniques to implement various Network Functions (NFs) like firewall, load balancer etc. from dedicated network devices to virtualized instances in commodity servers. This virtualized instance is called as Virtual Network Function (VNF). The purpose of VNF is to process NFs in order to accomplish a specific task. Traditionally these NFs were implemented on dedicated network equipment called middleboxes. Although these middleboxes are capable of processing heavy workloads, they are expensive, inflexible and require experts to maintain them. Therefore, NFV has the potential to substitute these middleboxes with virtualized instances in cloud datacenters, and hence greatly reduces the Operational Expenditure (OPEX) and Capital Expenditure (CAPEX) of networks by making it cheaper, flexible and scalable. This talk will share the recent advances and future challenges on how to build the infrastructure for hosting and managing NFVs. In particularly, we will focus on two of the most important topics in this research direction. First is the VNF placement problem, which aims to find the best mapping decision between VNF instances and physical resources. It has significant impact to the network operation cost, and service quality, but it is also known to be a NP-hard problem. So the problem has been actively studied by the research community. The second topic is Cloud-native/Container Network Function (CNF), which aims to minimize the overhead of traditional virtualization technique for network function using container-based technologies. Hence, it has drawn growing interests from industry to build the NFV infrastructure for CNF, but many new challenges remain to be addressed and studied.
今天的企业网络已经被称为网络功能虚拟化(NFV)的新兴技术彻底改变,NFV是欧洲电信标准协会(ETSI)提出的一种数据中心网络架构。NFV使用虚拟化技术实现各种网络功能(NFs),如防火墙、负载平衡器等,从专用网络设备到商用服务器中的虚拟化实例。这个虚拟实例称为VNF (Virtual Network Function)。VNF的目的是处理NFs,以完成特定的任务。传统上,这些NFs是在称为中间盒的专用网络设备上实现的。尽管这些中间设备能够处理繁重的工作负载,但它们价格昂贵、不灵活,需要专家来维护。因此,NFV有潜力用云数据中心中的虚拟化实例替代这些中间箱,从而通过使其更便宜、更灵活和可扩展,大大降低网络的运营支出(OPEX)和资本支出(CAPEX)。本次演讲将分享关于如何构建用于托管和管理nfv的基础设施的最新进展和未来挑战。特别是,我们将重点关注这一研究方向中最重要的两个主题。首先是VNF放置问题,其目的是找到VNF实例和物理资源之间的最佳映射决策。它对网络的运行成本和服务质量有很大的影响,同时也是一个np难题。因此,这个问题一直被研究界积极研究。第二个主题是云原生/容器网络功能(CNF),其目的是使用基于容器的技术将网络功能的传统虚拟化技术的开销降至最低。因此,为CNF构建NFV基础设施引起了业界越来越大的兴趣,但仍有许多新的挑战有待解决和研究。
{"title":"Recent Advances and Future Challenges for Network Function Virtualization Infrastructure","authors":"J. Chou","doi":"10.1145/3452411.3464449","DOIUrl":"https://doi.org/10.1145/3452411.3464449","url":null,"abstract":"Today's enterprise networks has revolutionized by the emerging technology called Network function virtualization (NFV), which is a type of data center network architecture proposed by the European Telecommunications Standards Institute (ETSI). NFV uses virtualization techniques to implement various Network Functions (NFs) like firewall, load balancer etc. from dedicated network devices to virtualized instances in commodity servers. This virtualized instance is called as Virtual Network Function (VNF). The purpose of VNF is to process NFs in order to accomplish a specific task. Traditionally these NFs were implemented on dedicated network equipment called middleboxes. Although these middleboxes are capable of processing heavy workloads, they are expensive, inflexible and require experts to maintain them. Therefore, NFV has the potential to substitute these middleboxes with virtualized instances in cloud datacenters, and hence greatly reduces the Operational Expenditure (OPEX) and Capital Expenditure (CAPEX) of networks by making it cheaper, flexible and scalable. This talk will share the recent advances and future challenges on how to build the infrastructure for hosting and managing NFVs. In particularly, we will focus on two of the most important topics in this research direction. First is the VNF placement problem, which aims to find the best mapping decision between VNF instances and physical resources. It has significant impact to the network operation cost, and service quality, but it is also known to be a NP-hard problem. So the problem has been actively studied by the research community. The second topic is Cloud-native/Container Network Function (CNF), which aims to minimize the overhead of traditional virtualization technique for network function using container-based technologies. Hence, it has drawn growing interests from industry to build the NFV infrastructure for CNF, but many new challenges remain to be addressed and studied.","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130134925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GPU-based Classification for Wireless Intrusion Detection 基于gpu的无线入侵检测分类
Pub Date : 2020-06-21 DOI: 10.1145/3452411.3464445
A. Lazar, A. Sim, Kesheng Wu
Automated network intrusion detection systems (NIDS) continuously monitor the network traffic to detect attacks or/and anomalies. These systems need to be able to detect attacks and alert network engineers in real-time. Therefore, modern NIDS are built using complex machine learning algorithms that require large training datasets and are time-consuming to train. The proposed work shows that machine learning algorithms from the RAPIDS cuML library on Graphics Processing Units (GPUs) can speed-up the training process on large scale datasets. This approach is able to reduce the training time while providing high accuracy and performance. We demonstrate the proposed approach on a large subset of data extracted from the Aegean Wi-Fi Intrusion Dataset (AWID). Multiple classification experiments were performed on both CPU and GPU. We achieve up to 65x acceleration of training several machine learning methods by moving most of the pipeline computations to the GPU and leveraging the new cuML library as well as the GPU version of the CatBoost library.
自动化网络入侵检测系统(NIDS)持续监控网络流量,以检测攻击或/或异常。这些系统需要能够检测攻击并实时提醒网络工程师。因此,现代NIDS是使用复杂的机器学习算法构建的,这些算法需要大量的训练数据集,并且训练起来很耗时。提出的工作表明,图形处理单元(gpu)上RAPIDS cuML库的机器学习算法可以加速大规模数据集的训练过程。这种方法能够在提供高精度和高性能的同时减少训练时间。我们在爱琴海Wi-Fi入侵数据集(AWID)中提取的大量数据上演示了所提出的方法。在CPU和GPU上分别进行了多次分类实验。通过将大部分管道计算移到GPU并利用新的cuML库以及GPU版本的CatBoost库,我们实现了几种机器学习方法的训练高达65倍的加速。
{"title":"GPU-based Classification for Wireless Intrusion Detection","authors":"A. Lazar, A. Sim, Kesheng Wu","doi":"10.1145/3452411.3464445","DOIUrl":"https://doi.org/10.1145/3452411.3464445","url":null,"abstract":"Automated network intrusion detection systems (NIDS) continuously monitor the network traffic to detect attacks or/and anomalies. These systems need to be able to detect attacks and alert network engineers in real-time. Therefore, modern NIDS are built using complex machine learning algorithms that require large training datasets and are time-consuming to train. The proposed work shows that machine learning algorithms from the RAPIDS cuML library on Graphics Processing Units (GPUs) can speed-up the training process on large scale datasets. This approach is able to reduce the training time while providing high accuracy and performance. We demonstrate the proposed approach on a large subset of data extracted from the Aegean Wi-Fi Intrusion Dataset (AWID). Multiple classification experiments were performed on both CPU and GPU. We achieve up to 65x acceleration of training several machine learning methods by moving most of the pipeline computations to the GPU and leveraging the new cuML library as well as the GPU version of the CatBoost library.","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130941321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Access Patterns to Disk Cache for Large Scientific Archive 大型科学档案磁盘缓存的访问模式
Pub Date : 2020-06-21 DOI: 10.1145/3452411.3464444
Yumeng Wang, Kesheng Wu, A. Sim, Shinjae Yoo, S. Misawa
Large scientific projects are increasing relying on analyses of data for their new discoveries; and a number of different data management systems have been developed to serve this scientific projects. In the work-in-progress paper, we describe an effort on understanding the data access patterns of one of these data management systems, dCache. This particular deployment of dCache acts as a disk cache in front of a large tape storage system primarily containing high-energy physics data. Based on the 15-month dCache logs, the cache is only accessing the tape system once for over 50 file requests, which indicates that it is effective as a disk cache. The on-disk files are repeated used, more than three times a day. We have also identified a number of unusual access patterns that are worth further investigation.
大型科学项目越来越依赖于对数据的分析来获得新发现;并且已经开发了许多不同的数据管理系统来服务于这些科学项目。在这篇正在进行的论文中,我们描述了理解这些数据管理系统之一dCache的数据访问模式的努力。dCache的这种特殊部署充当主要包含高能物理数据的大型磁带存储系统前面的磁盘缓存。根据15个月的dCache日志,当超过50个文件请求时,缓存只访问磁带系统一次,这表明它作为磁盘缓存是有效的。磁盘上的文件被重复使用,每天超过三次。我们还发现了一些不寻常的访问模式,值得进一步调查。
{"title":"Access Patterns to Disk Cache for Large Scientific Archive","authors":"Yumeng Wang, Kesheng Wu, A. Sim, Shinjae Yoo, S. Misawa","doi":"10.1145/3452411.3464444","DOIUrl":"https://doi.org/10.1145/3452411.3464444","url":null,"abstract":"Large scientific projects are increasing relying on analyses of data for their new discoveries; and a number of different data management systems have been developed to serve this scientific projects. In the work-in-progress paper, we describe an effort on understanding the data access patterns of one of these data management systems, dCache. This particular deployment of dCache acts as a disk cache in front of a large tape storage system primarily containing high-energy physics data. Based on the 15-month dCache logs, the cache is only accessing the tape system once for over 50 file requests, which indicates that it is effective as a disk cache. The on-disk files are repeated used, more than three times a day. We have also identified a number of unusual access patterns that are worth further investigation.","PeriodicalId":339207,"journal":{"name":"Proceedings of the 2021 on Systems and Network Telemetry and Analytics","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125052396","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Proceedings of the 2021 on Systems and Network Telemetry and Analytics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1