可扩展的,高速的基于片上的NDN名称转发使用FPGA

Divya Saxena, Suyash Mahar, V. Raychoudhury, Jiannong Cao
{"title":"可扩展的,高速的基于片上的NDN名称转发使用FPGA","authors":"Divya Saxena, Suyash Mahar, V. Raychoudhury, Jiannong Cao","doi":"10.1145/3288599.3288613","DOIUrl":null,"url":null,"abstract":"Named Data Networking (NDN) is the most promising candidate among the proposed content-based future Internet architectures. In NDN, Forwarding Information Base (FIB) maintains name prefixes and their corresponding outgoing interface(s) and forwards incoming packets by calculating the longest prefix match (LPM) of their content names (CNs). A CN in NDN is of variable-length and is maintained using a hierarchical structure. Therefore, to perform name lookup for packet forwarding at wire speed is a challenging task. However, the use of GPUs can lead to much better lookup speeds than CPU but, they are often limited by the CPU-GPU transfer latencies. In this paper, we exploit the massive parallel processing power of FPGA technology and propose a scalable, high-speed on-chip SRAM-based NDN name forwarding scheme for FIB (OnChip-FIB) using Field-Programmable Gate Arrays (FPGA). OnChip-FIB scales well as the number of prefixes grow, due to low storage complexity and low resource utilization. Extensive simulation results show that the OnChip-FIB scheme can achieve 1.06 μs measured lookup latency with a 26% on-chip block memory usage in a single Xilinx UltraScale FPGA for 50K named dataset.","PeriodicalId":346177,"journal":{"name":"Proceedings of the 20th International Conference on Distributed Computing and Networking","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2019-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Scalable, high-speed on-chip-based NDN name forwarding using FPGA\",\"authors\":\"Divya Saxena, Suyash Mahar, V. Raychoudhury, Jiannong Cao\",\"doi\":\"10.1145/3288599.3288613\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Named Data Networking (NDN) is the most promising candidate among the proposed content-based future Internet architectures. In NDN, Forwarding Information Base (FIB) maintains name prefixes and their corresponding outgoing interface(s) and forwards incoming packets by calculating the longest prefix match (LPM) of their content names (CNs). A CN in NDN is of variable-length and is maintained using a hierarchical structure. Therefore, to perform name lookup for packet forwarding at wire speed is a challenging task. However, the use of GPUs can lead to much better lookup speeds than CPU but, they are often limited by the CPU-GPU transfer latencies. In this paper, we exploit the massive parallel processing power of FPGA technology and propose a scalable, high-speed on-chip SRAM-based NDN name forwarding scheme for FIB (OnChip-FIB) using Field-Programmable Gate Arrays (FPGA). OnChip-FIB scales well as the number of prefixes grow, due to low storage complexity and low resource utilization. Extensive simulation results show that the OnChip-FIB scheme can achieve 1.06 μs measured lookup latency with a 26% on-chip block memory usage in a single Xilinx UltraScale FPGA for 50K named dataset.\",\"PeriodicalId\":346177,\"journal\":{\"name\":\"Proceedings of the 20th International Conference on Distributed Computing and Networking\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-01-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 20th International Conference on Distributed Computing and Networking\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3288599.3288613\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 20th International Conference on Distributed Computing and Networking","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3288599.3288613","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

命名数据网络(NDN)是提出的基于内容的未来互联网体系结构中最有前途的候选。在NDN中,FIB (Forwarding Information Base)维护名称前缀及其对应的出接口,并通过计算其内容名称(CNs)的LPM (longest prefix match)来转发进入的报文。NDN中的网络是可变长度的,并使用分层结构进行维护。因此,以线速执行包转发的名称查找是一项具有挑战性的任务。然而,使用gpu可以带来比CPU更好的查找速度,但是,它们通常受到CPU- gpu传输延迟的限制。在本文中,我们利用FPGA技术的巨大并行处理能力,提出了一个可扩展的,高速的基于片上sram的NDN名称转发方案,用于FIB (OnChip-FIB),使用现场可编程门阵列(FPGA)。由于存储复杂度低,资源利用率低,因此OnChip-FIB可以随着前缀数量的增加而扩展。大量的仿真结果表明,在单个Xilinx UltraScale FPGA上,对于50K命名数据集,OnChip-FIB方案可以在26%的片上块内存使用率下实现1.06 μs的测量查找延迟。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Scalable, high-speed on-chip-based NDN name forwarding using FPGA
Named Data Networking (NDN) is the most promising candidate among the proposed content-based future Internet architectures. In NDN, Forwarding Information Base (FIB) maintains name prefixes and their corresponding outgoing interface(s) and forwards incoming packets by calculating the longest prefix match (LPM) of their content names (CNs). A CN in NDN is of variable-length and is maintained using a hierarchical structure. Therefore, to perform name lookup for packet forwarding at wire speed is a challenging task. However, the use of GPUs can lead to much better lookup speeds than CPU but, they are often limited by the CPU-GPU transfer latencies. In this paper, we exploit the massive parallel processing power of FPGA technology and propose a scalable, high-speed on-chip SRAM-based NDN name forwarding scheme for FIB (OnChip-FIB) using Field-Programmable Gate Arrays (FPGA). OnChip-FIB scales well as the number of prefixes grow, due to low storage complexity and low resource utilization. Extensive simulation results show that the OnChip-FIB scheme can achieve 1.06 μs measured lookup latency with a 26% on-chip block memory usage in a single Xilinx UltraScale FPGA for 50K named dataset.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Improving efficacy of concurrent internal binary search trees using local recovery An accurate missing data prediction method using LSTM based deep learning for health care A simple and practical concurrent non-blocking unbounded graph with linearizable reachability queries EnTER: an encounter based trowbox deployment strategy for enhancing network reliability in post-disaster scenarios over DTN Exploration and impact of blockchain-enabled adaptive non-binary trust models
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1