首页 > 最新文献

2021 IEEE 29th International Conference on Network Protocols (ICNP)最新文献

英文 中文
DOVE: Diagnosis-driven SLO Violation Detection DOVE:诊断驱动的SLO违规检测
Pub Date : 2021-11-01 DOI: 10.1109/ICNP52444.2021.9651986
Yiran Lei, Yu Zhou, Yunsenxiao Lin, Mingwei Xu, Yangyang Wang
Service-level objectives (SLOs), as network performance requirements for delay and packet loss typically, should be guaranteed for increasing high-performance applications, e.g., telesurgery and cloud gaming. However, SLO violations are common and destructive in today’s network operation. Detection and diagnosis, meaning monitoring performance to discover anomalies and analyzing causality of SLO violations respectively, are crucial for fast recovery. Unfortunately, existing diagnosis approaches require exhaustive causal information to function. Meanwhile, existing detection tools incur large overhead or are only able to provide limited information for diagnosis. This paper presents DOVE, a diagnosis-driven SLO detection system with high accuracy and low overhead. The key idea is to identify and report the information needed by diagnosis along with SLO violation alerts from the data plane selectively and efficiently. Network segmentation is introduced to balance scalability and accuracy. Novel algorithms to measure packet loss and percentile delay are implemented completely on the data plane without the involvement of the control plane for fine-grained SLO detection. We implement and deploy DOVE on Tofino and P4 software switch (BMv2) and show the effectiveness of DOVE with a use case. The reported SLO violation alerts and diagnosis-needing information are compared with ground truth and show high accuracy (>97%). Our evaluation also shows that DOVE introduces up to two orders of magnitude less traffic overhead than NetSight. In addition, memory utilization and required processing ability are low to be deployable in real network topologies.
服务水平目标(slo),作为延迟和数据包丢失的网络性能要求,应该保证不断增长的高性能应用程序,例如远程外科手术和云游戏。然而,在当今的网络运营中,违反SLO是常见的和具有破坏性的。检测和诊断,即分别监控性能以发现异常和分析违反SLO的因果关系,对于快速恢复至关重要。不幸的是,现有的诊断方法需要详尽的因果信息才能发挥作用。同时,现有的检测工具开销较大,或者只能提供有限的诊断信息。本文介绍了一种诊断驱动的高精度低开销SLO检测系统DOVE。关键思想是有选择地有效地识别和报告诊断所需的信息以及来自数据平面的SLO违规警报。为了平衡可扩展性和准确性,引入了网络分段。为了实现细粒度的SLO检测,在数据平面上完全实现了测量丢包和百分位延迟的新算法,而不需要控制平面的参与。我们在Tofino和P4软件交换机(BMv2)上实现和部署了DOVE,并通过一个用例展示了DOVE的有效性。将报告的SLO违规警报和需要诊断的信息与实际情况进行比较,显示出较高的准确性(>97%)。我们的评估还表明,DOVE带来的流量开销比NetSight少两个数量级。此外,内存利用率和所需的处理能力较低,无法在实际网络拓扑中部署。
{"title":"DOVE: Diagnosis-driven SLO Violation Detection","authors":"Yiran Lei, Yu Zhou, Yunsenxiao Lin, Mingwei Xu, Yangyang Wang","doi":"10.1109/ICNP52444.2021.9651986","DOIUrl":"https://doi.org/10.1109/ICNP52444.2021.9651986","url":null,"abstract":"Service-level objectives (SLOs), as network performance requirements for delay and packet loss typically, should be guaranteed for increasing high-performance applications, e.g., telesurgery and cloud gaming. However, SLO violations are common and destructive in today’s network operation. Detection and diagnosis, meaning monitoring performance to discover anomalies and analyzing causality of SLO violations respectively, are crucial for fast recovery. Unfortunately, existing diagnosis approaches require exhaustive causal information to function. Meanwhile, existing detection tools incur large overhead or are only able to provide limited information for diagnosis. This paper presents DOVE, a diagnosis-driven SLO detection system with high accuracy and low overhead. The key idea is to identify and report the information needed by diagnosis along with SLO violation alerts from the data plane selectively and efficiently. Network segmentation is introduced to balance scalability and accuracy. Novel algorithms to measure packet loss and percentile delay are implemented completely on the data plane without the involvement of the control plane for fine-grained SLO detection. We implement and deploy DOVE on Tofino and P4 software switch (BMv2) and show the effectiveness of DOVE with a use case. The reported SLO violation alerts and diagnosis-needing information are compared with ground truth and show high accuracy (>97%). Our evaluation also shows that DOVE introduces up to two orders of magnitude less traffic overhead than NetSight. In addition, memory utilization and required processing ability are low to be deployable in real network topologies.","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125869293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Welcome Message from the ICNP 2021 TPC Chairs ICNP 2021 TPC主席欢迎辞
Pub Date : 2021-11-01 DOI: 10.1109/icnp52444.2021.9651933
{"title":"Welcome Message from the ICNP 2021 TPC Chairs","authors":"","doi":"10.1109/icnp52444.2021.9651933","DOIUrl":"https://doi.org/10.1109/icnp52444.2021.9651933","url":null,"abstract":"","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130147487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Poster: Accelerate Cross-Device Federated Learning With Semi-Reliable Model Multicast Over The Air 海报:用半可靠的空中多播模型加速跨设备联邦学习
Pub Date : 2021-11-01 DOI: 10.1109/ICNP52444.2021.9651964
Yunzhi Lin, Shouxi Luo
To achieve efficient model multicast for cross-device Federated Learning (FL) over shared wireless channels, we propose SRMP, a transport protocol that performs semi-reliable model multicast over the air by leveraging existing PHY-aided wireless multicast techniques. The preliminary study shows that, with novel designs, SRMP could reduce the communication time involved in each round of training significantly.
为了在共享无线信道上实现跨设备联邦学习(FL)的高效模型组播,我们提出了SRMP,这是一种利用现有的物理辅助无线组播技术在空中执行半可靠模型组播的传输协议。初步研究表明,通过新颖的设计,SRMP可以显著减少每轮训练的沟通时间。
{"title":"Poster: Accelerate Cross-Device Federated Learning With Semi-Reliable Model Multicast Over The Air","authors":"Yunzhi Lin, Shouxi Luo","doi":"10.1109/ICNP52444.2021.9651964","DOIUrl":"https://doi.org/10.1109/ICNP52444.2021.9651964","url":null,"abstract":"To achieve efficient model multicast for cross-device Federated Learning (FL) over shared wireless channels, we propose SRMP, a transport protocol that performs semi-reliable model multicast over the air by leveraging existing PHY-aided wireless multicast techniques. The preliminary study shows that, with novel designs, SRMP could reduce the communication time involved in each round of training significantly.","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116490447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Loss-freedom, Order-preservation and No-buffering: Pick Any Two During Flow Migration in Network Functions 网络函数流迁移过程中无损失、保序和无缓冲:任选其一
Pub Date : 2021-11-01 DOI: 10.1109/ICNP52444.2021.9651954
Radhika Sukapuram, Ranjan Patowary, G. Barua
Network Functions (NFs) provide security and optimization services to networks by examining and modifying packets and by collecting information. When NFs need to be scaled out to manage higher load or scaled in to conserve energy, flows need to be migrated from one instance of an NF, called the source instance, to another, called the destination instance, or from one chain of instances to another chain of instances. Before flows are migrated, the state information associated with the source instance needs to be migrated to the destination instance. Packets that arrive at the destination instance meanwhile need to be either buffered or dropped until the state information is migrated, for correct functioning of some stateful NFs, while for some others, the destination NF may continue to function. We define the properties of Loss-freedom, where the flow migration system does not drop packets, No-buffering, where it does not buffer packets, and Order-preservation, where it processes packets in the same manner as the source NF, if there was no flow migration. We formalize these properties, for the first time, and prove that it is impossible for a flow migration algorithm in stateful NFs to guarantee satisfying all three of the properties of Loss-freedom (L), Order-preservation (O) and No-buffering (N) during flow migration, even if messages or packets are not lost. We demonstrate how existing algorithms operate with regard to these properties and prove that these properties are compositional.
NFs (Network Functions)通过对报文的检测、修改和信息收集等方式,为网络提供安全和优化服务。当NFs需要向外扩展以管理更高的负载或向内扩展以节省能源时,需要将流从一个NF实例(称为源实例)迁移到另一个NF实例(称为目标实例),或者从一个实例链迁移到另一个实例链。在迁移流之前,需要将与源实例关联的状态信息迁移到目标实例。同时到达目标实例的数据包需要被缓冲或丢弃,直到状态信息被迁移,以便一些有状态NFs的正常运行,而对于其他一些,目标NF可能继续运行。我们定义了Loss-freedom属性,即流迁移系统不丢包;no -buffering属性,即不缓冲数据包;order - preserving属性,即在没有流迁移的情况下,它以与源NF相同的方式处理数据包。我们首次形式化了这些性质,并证明了在有状态NFs中的流迁移算法不可能保证在流迁移过程中满足所有三个性质:丢失自由(L)、顺序保持(O)和无缓冲(N),即使消息或数据包没有丢失。我们演示了现有的算法是如何处理这些属性的,并证明了这些属性是组合的。
{"title":"Loss-freedom, Order-preservation and No-buffering: Pick Any Two During Flow Migration in Network Functions","authors":"Radhika Sukapuram, Ranjan Patowary, G. Barua","doi":"10.1109/ICNP52444.2021.9651954","DOIUrl":"https://doi.org/10.1109/ICNP52444.2021.9651954","url":null,"abstract":"Network Functions (NFs) provide security and optimization services to networks by examining and modifying packets and by collecting information. When NFs need to be scaled out to manage higher load or scaled in to conserve energy, flows need to be migrated from one instance of an NF, called the source instance, to another, called the destination instance, or from one chain of instances to another chain of instances. Before flows are migrated, the state information associated with the source instance needs to be migrated to the destination instance. Packets that arrive at the destination instance meanwhile need to be either buffered or dropped until the state information is migrated, for correct functioning of some stateful NFs, while for some others, the destination NF may continue to function. We define the properties of Loss-freedom, where the flow migration system does not drop packets, No-buffering, where it does not buffer packets, and Order-preservation, where it processes packets in the same manner as the source NF, if there was no flow migration. We formalize these properties, for the first time, and prove that it is impossible for a flow migration algorithm in stateful NFs to guarantee satisfying all three of the properties of Loss-freedom (L), Order-preservation (O) and No-buffering (N) during flow migration, even if messages or packets are not lost. We demonstrate how existing algorithms operate with regard to these properties and prove that these properties are compositional.","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126714577","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Demo: Simple Deep Packet Inspection with P4 演示:简单的深度包检测与P4
Pub Date : 2021-11-01 DOI: 10.1109/ICNP52444.2021.9651973
Sahil Gupta, D. Gosain, Garegin Grigoryan, Minseok Kwon, H. B. Acharya
The P4 language allows "protocol-independent packet parsing" in network switches, and makes many operations possible in the data plane. But P4 is not built for Deep Packet Inspection – it can only "parse" well-defined packet headers, not free-form headers as seen in HTTPS etc. Thus some very important use cases, such as application-layer firewalls, are considered impossible for P4. This demonstration shows that this limitation is not strictly true: switches, that support only standard P4, are able to independently perform tasks such as blocking specific URLs (without using non-standard "extern" components, help from the SDN controller, or rerouting to a firewall). As more Internet infrastructure becomes SDN-compatible, in future, switches may perform simple application-layer firewall tasks.
P4语言允许在网络交换机中“独立于协议的数据包解析”,并使数据平面上的许多操作成为可能。但是P4不是为深度包检测而构建的——它只能“解析”定义良好的包头,而不是像HTTPS等中看到的自由格式的头。因此,一些非常重要的用例,如应用层防火墙,被认为不可能用于P4。这个演示表明,这种限制并不完全正确:仅支持标准P4的交换机能够独立执行诸如阻止特定url之类的任务(无需使用非标准的“外部”组件、SDN控制器的帮助或重路由到防火墙)。随着越来越多的Internet基础设施与sdn兼容,将来交换机可能会执行简单的应用层防火墙任务。
{"title":"Demo: Simple Deep Packet Inspection with P4","authors":"Sahil Gupta, D. Gosain, Garegin Grigoryan, Minseok Kwon, H. B. Acharya","doi":"10.1109/ICNP52444.2021.9651973","DOIUrl":"https://doi.org/10.1109/ICNP52444.2021.9651973","url":null,"abstract":"The P4 language allows \"protocol-independent packet parsing\" in network switches, and makes many operations possible in the data plane. But P4 is not built for Deep Packet Inspection – it can only \"parse\" well-defined packet headers, not free-form headers as seen in HTTPS etc. Thus some very important use cases, such as application-layer firewalls, are considered impossible for P4. This demonstration shows that this limitation is not strictly true: switches, that support only standard P4, are able to independently perform tasks such as blocking specific URLs (without using non-standard \"extern\" components, help from the SDN controller, or rerouting to a firewall). As more Internet infrastructure becomes SDN-compatible, in future, switches may perform simple application-layer firewall tasks.","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127752279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
StaR: Breaking the Scalability Limit for RDMA StaR:打破RDMA的可扩展性限制
Pub Date : 2021-11-01 DOI: 10.1109/ICNP52444.2021.9651935
Xizheng Wang, Guo Chen, Xijin Yin, Huichen Dai, Bojie Li, Binzhang Fu, Kun Tan
Due to its superior performance, Remote Direct Memory Access (RDMA) has been widely deployed in data center networks. It provides applications with ultra-high throughput, ultra-low latency, and far lower CPU utilization than TCP/IP software network stack. However, the connection states that must be stored on the RDMA NIC (RNIC) and the small NIC memory result in poor scalability. The performance drops significantly when the RNIC needs to maintain a large number of concurrent connections.We propose StaR (Stateless RDMA), which solves the scalability problem of RDMA by transferring states to the other communication end. Leveraging the asymmetric communication pattern in data center applications, StaR lets the communication end with low concurrency save states for the other end with high concurrency, thus making the RNIC on the bottleneck side to be stateless. We have implemented StaR on an FPGA board with 10Gbps network port and evaluated its performance on a testbed with 9 machines all equipped with StaR NICs. The experimental results show that in high concurrency scenarios, the throughput of StaR can reach up to 4.13x and 1.35x of the original RNIC and the latest software-based solution, respectively.
RDMA (Remote Direct Memory Access)由于其优越的性能,在数据中心网络中得到了广泛的应用。它为应用程序提供了比TCP/IP软件网络堆栈更高的吞吐量、更低的延迟和更低的CPU利用率。但是,必须存储在RDMA网卡(RNIC)上的连接状态和较小的网卡内存导致可扩展性较差。当RNIC需要维护大量并发连接时,性能会明显下降。我们提出了StaR(无状态RDMA),它通过向另一端传输状态来解决RDMA的可扩展性问题。利用数据中心应用程序中的非对称通信模式,StaR允许具有低并发性的通信端为具有高并发性的另一端保存状态,从而使瓶颈端的RNIC处于无状态状态。我们在带有10Gbps网络端口的FPGA板上实现了StaR,并在配备StaR网卡的9台机器的测试台上对其性能进行了评估。实验结果表明,在高并发场景下,StaR的吞吐量分别可以达到原始RNIC和最新基于软件的解决方案的4.13倍和1.35倍。
{"title":"StaR: Breaking the Scalability Limit for RDMA","authors":"Xizheng Wang, Guo Chen, Xijin Yin, Huichen Dai, Bojie Li, Binzhang Fu, Kun Tan","doi":"10.1109/ICNP52444.2021.9651935","DOIUrl":"https://doi.org/10.1109/ICNP52444.2021.9651935","url":null,"abstract":"Due to its superior performance, Remote Direct Memory Access (RDMA) has been widely deployed in data center networks. It provides applications with ultra-high throughput, ultra-low latency, and far lower CPU utilization than TCP/IP software network stack. However, the connection states that must be stored on the RDMA NIC (RNIC) and the small NIC memory result in poor scalability. The performance drops significantly when the RNIC needs to maintain a large number of concurrent connections.We propose StaR (Stateless RDMA), which solves the scalability problem of RDMA by transferring states to the other communication end. Leveraging the asymmetric communication pattern in data center applications, StaR lets the communication end with low concurrency save states for the other end with high concurrency, thus making the RNIC on the bottleneck side to be stateless. We have implemented StaR on an FPGA board with 10Gbps network port and evaluated its performance on a testbed with 9 machines all equipped with StaR NICs. The experimental results show that in high concurrency scenarios, the throughput of StaR can reach up to 4.13x and 1.35x of the original RNIC and the latest software-based solution, respectively.","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"66 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133609501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Hopping on Spectrum: Measuring and Boosting a Large-scale Dual-band Wireless Network 频谱上的跳跃:测量和推进大规模双频无线网络
Pub Date : 2021-11-01 DOI: 10.1109/ICNP52444.2021.9651921
Haibo Wang, Tao Gao, Weizhen Dang, Jing’an Xue, Jiahao Cao, Fenghua Li, Jilong Wang
In recent years, more and more wireless networks support both 2.4GHz and 5GHz bands. However, in large-scale dual-band wireless networks, lack of understanding on the behavior and performance makes the network diagnosis and optimization extremely challenging. In this paper, we conduct a comprehensive measurement to characterize the behavior and performance in a large-scale dual-band wireless network (TD WLAN). We make several meaningful observations. (1) Although the 5GHz band outperforms the 2.4GHz band, 60% of devices tend to be associated with the 2.4GHz band. The device association behavior has a large impact on the performance. (2) Rogue and non-WiFi devices are prevalent, wherein hidden terminal interference increases the average loss rate by 8%, carrier sense interference increases the average WiFi latency by 45%, and RF interference further aggravates both packet loss and channel contention. (3) The dynamic channel assignment strategy is not always effective. On this basis, we propose a novel and easy-to-implement strategy to improve the wireless performance by intelligent band navigation and heuristic channel optimization. The actual deployment in TD WLAN shows the packet loss reduces by 40% on average and the WiFi latency for more than 60% of devices is below 5ms.
近年来,越来越多的无线网络同时支持2.4GHz和5GHz频段。然而,在大规模双频无线网络中,由于缺乏对网络行为和性能的认识,使得网络的诊断和优化工作极具挑战性。在本文中,我们进行了全面的测量,以表征大规模双频无线网络(TD WLAN)的行为和性能。我们做了一些有意义的观察。(1) 5GHz频段虽然优于2.4GHz频段,但60%的设备倾向于与2.4GHz频段关联。设备关联行为对性能影响较大。(2)恶意和非WiFi设备普遍存在,其中隐性终端干扰使平均丢包率增加8%,载波感知干扰使平均WiFi延迟增加45%,射频干扰进一步加剧丢包和信道争用。(3)动态信道分配策略并不总是有效的。在此基础上,我们提出了一种新颖且易于实现的策略,通过智能频段导航和启发式信道优化来提高无线性能。在TD WLAN中的实际部署表明,丢包率平均降低了40%,超过60%的设备的WiFi延迟低于5ms。
{"title":"Hopping on Spectrum: Measuring and Boosting a Large-scale Dual-band Wireless Network","authors":"Haibo Wang, Tao Gao, Weizhen Dang, Jing’an Xue, Jiahao Cao, Fenghua Li, Jilong Wang","doi":"10.1109/ICNP52444.2021.9651921","DOIUrl":"https://doi.org/10.1109/ICNP52444.2021.9651921","url":null,"abstract":"In recent years, more and more wireless networks support both 2.4GHz and 5GHz bands. However, in large-scale dual-band wireless networks, lack of understanding on the behavior and performance makes the network diagnosis and optimization extremely challenging. In this paper, we conduct a comprehensive measurement to characterize the behavior and performance in a large-scale dual-band wireless network (TD WLAN). We make several meaningful observations. (1) Although the 5GHz band outperforms the 2.4GHz band, 60% of devices tend to be associated with the 2.4GHz band. The device association behavior has a large impact on the performance. (2) Rogue and non-WiFi devices are prevalent, wherein hidden terminal interference increases the average loss rate by 8%, carrier sense interference increases the average WiFi latency by 45%, and RF interference further aggravates both packet loss and channel contention. (3) The dynamic channel assignment strategy is not always effective. On this basis, we propose a novel and easy-to-implement strategy to improve the wireless performance by intelligent band navigation and heuristic channel optimization. The actual deployment in TD WLAN shows the packet loss reduces by 40% on average and the WiFi latency for more than 60% of devices is below 5ms.","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133279193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MagicTCAM: A Multiple-TCAM Scheme for Fast TCAM Update MagicTCAM:一个快速TCAM更新的多TCAM方案
Pub Date : 2021-11-01 DOI: 10.1109/ICNP52444.2021.9651965
Ruyi Yao, Cong Luo, Xuandong Liu, Ying Wan, B. Liu, Wen J. Li, Yang Xu
Ternary Content-Addressable Memory (TCAM) is a popular solution for high-speed flow table lookup in Software-Defined Networking (SDN). Rule insertion in TCAM is a time-consuming operation. To ensure semantic correctness, rules overlapped must be stored in TCAM with decreasing priority order and many rule movements may be needed to make space for a single inserted rule. When a rule insertion is in progress, the regular flow table lookup will be suspended, which could lead to a degraded user experience for SDN applications. In this paper, we propose a multiple-TCAM framework named MagicTCAM to reduce the rule movements during a rule insertion. The core of MagicTCAM lies in three operations: layering, partitioning and rotating. By layering, rules with the least overlapping will be grouped (i.e., layered) into a sub-ruleset. The number of rule movements is therefore greatly reduced as most of rules in a sub-ruleset are non-overlapped. To achieve balanced load in TCAMs, rules in each sub-ruleset are further partitioned and dispatched into different TCAMs in a rotating manner. In addition, an inter-TCAM movement algorithm is proposed to allow rules to be moved between TCAMs for reduced rule movement. Experiment results show that with two half-sized TCAMs, MagicTCAM reduces the rule movements by 39% on average compared with the state-of-the-art work while the computation time is shortened by half as well.
三元内容可寻址存储器(TCAM)是软件定义网络(SDN)中用于高速流表查找的一种流行解决方案。在TCAM中插入规则是一项耗时的操作。为了确保语义的正确性,必须将重叠的规则按优先级递减的顺序存储在TCAM中,并且可能需要许多规则移动来为单个插入的规则腾出空间。当进行规则插入时,将暂停常规流表查找,这可能导致SDN应用程序的用户体验下降。在本文中,我们提出了一个名为MagicTCAM的多tcam框架来减少规则插入过程中的规则移动。MagicTCAM的核心在于三个操作:分层、分区和旋转。通过分层,重叠最少的规则将被分组(即分层)到子规则集中。因此,规则移动的数量大大减少,因为子规则集中的大多数规则是不重叠的。为了实现tcam中的负载均衡,将每个子规则集中的规则进一步划分,并以旋转的方式分配到不同的tcam中。此外,提出了一种跨tcam移动算法,允许规则在tcam之间移动,以减少规则的移动。实验结果表明,对于两个半尺寸的tcam, MagicTCAM与目前的工作相比,平均减少了39%的规则运动,计算时间也缩短了一半。
{"title":"MagicTCAM: A Multiple-TCAM Scheme for Fast TCAM Update","authors":"Ruyi Yao, Cong Luo, Xuandong Liu, Ying Wan, B. Liu, Wen J. Li, Yang Xu","doi":"10.1109/ICNP52444.2021.9651965","DOIUrl":"https://doi.org/10.1109/ICNP52444.2021.9651965","url":null,"abstract":"Ternary Content-Addressable Memory (TCAM) is a popular solution for high-speed flow table lookup in Software-Defined Networking (SDN). Rule insertion in TCAM is a time-consuming operation. To ensure semantic correctness, rules overlapped must be stored in TCAM with decreasing priority order and many rule movements may be needed to make space for a single inserted rule. When a rule insertion is in progress, the regular flow table lookup will be suspended, which could lead to a degraded user experience for SDN applications. In this paper, we propose a multiple-TCAM framework named MagicTCAM to reduce the rule movements during a rule insertion. The core of MagicTCAM lies in three operations: layering, partitioning and rotating. By layering, rules with the least overlapping will be grouped (i.e., layered) into a sub-ruleset. The number of rule movements is therefore greatly reduced as most of rules in a sub-ruleset are non-overlapped. To achieve balanced load in TCAMs, rules in each sub-ruleset are further partitioned and dispatched into different TCAMs in a rotating manner. In addition, an inter-TCAM movement algorithm is proposed to allow rules to be moved between TCAMs for reduced rule movement. Experiment results show that with two half-sized TCAMs, MagicTCAM reduces the rule movements by 39% on average compared with the state-of-the-art work while the computation time is shortened by half as well.","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124765052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Receiver-Driven RDMA Congestion Control by Differentiating Congestion Types in Datacenter Networks 通过区分数据中心网络中的拥塞类型来控制接收端驱动的RDMA拥塞
Pub Date : 2021-11-01 DOI: 10.1109/ICNP52444.2021.9651938
Jiao Zhang, Jiaming Shi, Xiaolong Zhong, Zirui Wan, Yuxing Tian, Tian Pan, Tao Huang
The development of datacenter applications leads to the need for end-to-end communication with microsecond latency. As a result, RDMA is becoming prevalent in datacenter networks to mitigate the latency caused by the slow processing speed of the traditional software network stack. However, existing RDMA congestion control mechanisms are either far from optimal in simultaneously achieving high throughput and low latency or in need of additional in-network function support. In this paper, by leveraging the observation that most congestion occurs at the last hop in datacenter networks, we propose RCC, a receiver-driven rapid congestion control mechanism for RDMA networks that combines explicit assignment and iterative window adjustment. Firstly, we propose a network congestion distinguish method to classify congestions into two types, last-hop congestion and innetwork congestion. Then, an Explicit Window Assignment mechanism is proposed to solve the last-hop congestion, which enables senders to converge to a proper sending rate in one-RTT. For in-network congestion, a PID-based iterative delay-based window adjustment scheme is proposed to achieve fast convergence and near-zero queuing latency. RCC does not need additional innetwork support and is friendly to hardware implementation. In our evaluation, the overall average FCT (Flow Completion Time) of RCC is 4~79% better than Homa, ExpressPass, DCQCN, TIMELY, and HPCC.
数据中心应用程序的开发导致需要具有微秒级延迟的端到端通信。因此,RDMA在数据中心网络中变得越来越普遍,以减轻传统软件网络堆栈处理速度慢所造成的延迟。然而,现有的RDMA拥塞控制机制在同时实现高吞吐量和低延迟方面远非最佳,或者需要额外的网络内功能支持。在本文中,通过观察大多数拥塞发生在数据中心网络的最后一跳,我们提出了RCC,一种接收器驱动的RDMA网络快速拥塞控制机制,结合了显式分配和迭代窗口调整。首先,提出了一种网络拥塞判别方法,将拥塞分为最后一跳拥塞和网络内拥塞两种类型。然后,提出了一种显式窗口分配机制来解决最后一跳拥塞问题,使发送端能够在一次rtt中收敛到合适的发送速率。针对网络内拥塞问题,提出了一种基于pid的迭代延迟窗口调整方案,实现了快速收敛和接近于零的排队延迟。RCC不需要额外的网络支持,并且对硬件实现很友好。在我们的评估中,RCC的总体平均FCT (Flow Completion Time)比Homa、ExpressPass、DCQCN、TIMELY和HPCC好4~79%。
{"title":"Receiver-Driven RDMA Congestion Control by Differentiating Congestion Types in Datacenter Networks","authors":"Jiao Zhang, Jiaming Shi, Xiaolong Zhong, Zirui Wan, Yuxing Tian, Tian Pan, Tao Huang","doi":"10.1109/ICNP52444.2021.9651938","DOIUrl":"https://doi.org/10.1109/ICNP52444.2021.9651938","url":null,"abstract":"The development of datacenter applications leads to the need for end-to-end communication with microsecond latency. As a result, RDMA is becoming prevalent in datacenter networks to mitigate the latency caused by the slow processing speed of the traditional software network stack. However, existing RDMA congestion control mechanisms are either far from optimal in simultaneously achieving high throughput and low latency or in need of additional in-network function support. In this paper, by leveraging the observation that most congestion occurs at the last hop in datacenter networks, we propose RCC, a receiver-driven rapid congestion control mechanism for RDMA networks that combines explicit assignment and iterative window adjustment. Firstly, we propose a network congestion distinguish method to classify congestions into two types, last-hop congestion and innetwork congestion. Then, an Explicit Window Assignment mechanism is proposed to solve the last-hop congestion, which enables senders to converge to a proper sending rate in one-RTT. For in-network congestion, a PID-based iterative delay-based window adjustment scheme is proposed to achieve fast convergence and near-zero queuing latency. RCC does not need additional innetwork support and is friendly to hardware implementation. In our evaluation, the overall average FCT (Flow Completion Time) of RCC is 4~79% better than Homa, ExpressPass, DCQCN, TIMELY, and HPCC.","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124553827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
MPLibra: Complementing the Benefits of Classic and Learning-based Multipath Congestion Control MPLibra:补充经典和基于学习的多路径拥塞控制的优势
Pub Date : 2021-11-01 DOI: 10.1109/ICNP52444.2021.9651987
Hebin Yu, Jiaqi Zheng, Zhuoxuan Du, Guihai Chen
Multipath TCP (MPTCP) is a burgeoning transport protocol which enables the server to split the traffic across multiple network interfaces. Classic MPTCPs have good friendliness and practicality such as relatively low overhead, but are hard to achieve consistent high-throughput and adaptability, especially for the ability of flexibly balancing congestion among different paths. In contrast, learning-based MPTCPs can essentially achieve consistent high-throughput and adaptability, but have poor friendliness and practicality. In this paper, we proposed MPLibra, a combined multipath congestion control framework that can complement the advantages of classic MPTCPs and learning-based MPTCPs. Extensive simulations on NS3 show that MPLibra can achieve good performance and outperform state-of-the-art MPTCPs under different network conditions. MPLibra improves the throughput by 40.5% and reduces the file download time by 47.7% compared with LIA, achieves good friendliness and balances congestion timely.
多路径TCP (MPTCP)是一种新兴的传输协议,它使服务器能够跨多个网络接口拆分流量。传统的mptcp具有良好的友好性和实用性,开销相对较低,但难以实现一致的高吞吐量和适应性,特别是在不同路径间灵活平衡拥塞的能力方面。而基于学习的mptcp基本上可以实现一致的高吞吐量和适应性,但友好性和实用性较差。在本文中,我们提出了MPLibra,这是一个组合的多路径拥塞控制框架,可以补充经典mptcp和基于学习的mptcp的优点。在NS3上的大量仿真表明,MPLibra可以在不同的网络条件下获得良好的性能,并且优于最先进的mptcp。与LIA相比,MPLibra提高了40.5%的吞吐量,减少了47.7%的文件下载时间,实现了良好的友好性,及时平衡了拥塞。
{"title":"MPLibra: Complementing the Benefits of Classic and Learning-based Multipath Congestion Control","authors":"Hebin Yu, Jiaqi Zheng, Zhuoxuan Du, Guihai Chen","doi":"10.1109/ICNP52444.2021.9651987","DOIUrl":"https://doi.org/10.1109/ICNP52444.2021.9651987","url":null,"abstract":"Multipath TCP (MPTCP) is a burgeoning transport protocol which enables the server to split the traffic across multiple network interfaces. Classic MPTCPs have good friendliness and practicality such as relatively low overhead, but are hard to achieve consistent high-throughput and adaptability, especially for the ability of flexibly balancing congestion among different paths. In contrast, learning-based MPTCPs can essentially achieve consistent high-throughput and adaptability, but have poor friendliness and practicality. In this paper, we proposed MPLibra, a combined multipath congestion control framework that can complement the advantages of classic MPTCPs and learning-based MPTCPs. Extensive simulations on NS3 show that MPLibra can achieve good performance and outperform state-of-the-art MPTCPs under different network conditions. MPLibra improves the throughput by 40.5% and reduces the file download time by 47.7% compared with LIA, achieves good friendliness and balances congestion timely.","PeriodicalId":343813,"journal":{"name":"2021 IEEE 29th International Conference on Network Protocols (ICNP)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117119809","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2021 IEEE 29th International Conference on Network Protocols (ICNP)
全部 Geostand. Geoanal. Res. Nat. Geosci. 环境与发展 Archaeol. Anthropol. Sci. High Temp. Commun. Phys. Environ. Eng. Sci. Atmos. Chem. Phys. J. Atmos. Oceanic Technol. 2011 International Conference on Electric Technology and Civil Engineering (ICETCE) Asia-Pac. J. Atmos. Sci. Environ. Eng. Manage. J. Mineral. Mag. J. Earth Sci. EUR PHYS J-APPL PHYS Fossil Rec. Curr. Appl Phys. 2011 International Conference on Computer Distributed Control and Intelligent Environmental Monitoring Addict. Behav. Carbon Balance Manage. TECTONOPHYSICS BRAZ J INFECT DIS Environ. Technol. Innovation Environ. Prog. Sustainable Energy Acta Neuropsychiatr. Toung Pao J. Synchrotron Radiat. Ann. Glaciol. 1997 IEEE Ultrasonics Symposium Proceedings. An International Symposium (Cat. No.97CH36118) Condens. Matter Phys. Chem. Ecol. 2008 Annual Report Conference on Electrical Insulation and Dielectric Phenomena J APPL METEOROL CLIM ARCHAEOMETRY Global Biogeochem. Cycles Energy Environ. Acta Geochimica Mod. Phys. Lett. A Environmental dermatology : the official journal of the Japanese Society for Contact Dermatitis [Sanfujinka chiryo] Obstetrical and gynecological therapy ACTA DIABETOL BIOGEOSCIENCES Commun. Theor. Phys. Acta Oceanolog. Sin. Swiss J. Geosci. WEATHER Quat. Res. J. Atmos. Sol. Terr. Phys. OCEAN SCI J COMP BIOCHEM PHYS C Nat. Clim. Change ACTA GEOL SIN-ENGL Appl. Geochem. Am. J. Phys. Anthropol. J. Atmos. Chem. Aquat. Geochem. IZV-PHYS SOLID EART+ Adv. Meteorol. Acta Geophys. Org. Geochem. Aust. J. Earth Sci. Geochim. Cosmochim. Acta J. Hydrol. ACTA PETROL SIN Int. J. Biometeorol. Am. Mineral. Annu. Rev. Earth Planet. Sci. Geobiology Contrib. Mineral. Petrol. AAPG Bull. Appl. Clay Sci. Geochem. J. Adv. Atmos. Sci. CHIN OPT LETT ACTA GEOL POL Geochem. Trans. Clean Technol. Environ. Policy ARCT ANTARCT ALP RES J. Geog. Sci. Atmos. Res. Geol. Ore Deposits Basin Res. Geosci. J. ECOLOGY Conserv. Genet. Resour. IEEE Trans. Appl. Supercond. Can. J. Phys. J. Appl. Phys. ACTA ONCOL Chin. Phys. C Isl. Arc EUR PHYS J-SPEC TOP Clean-Soil Air Water Ecol. Res. Int. Geol. Rev. Conserv. Biol. Geochem. Int. Big Earth Data Geosci. Front. ACTA PARASITOL
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1