首页 > 最新文献

Proceedings of the 16th ACM Workshop on Hot Topics in Networks最新文献

英文 中文
DIY Hosting for Online Privacy DIY在线隐私托管
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152459
Shoumik Palkar, M. Zaharia
Web users today rely on centralized services for applications such as email, file transfer and chat. Unfortunately, these services create a significant privacy risk: even with a benevolent provider, a single breach can put millions of users' data at risk. One alternative would be for users to host their own servers, but this would be highly expensive for most applications: a single VM deployed in a high-availability mode can cost many dollars per month. In this paper, we propose Deploy It Yourself (DIY), a new model for hosting applications based on serverless computing platforms such as Amazon Lambda. DIY allows users to run a highly available service with much stronger privacy guarantees than current centralized providers, and at a dramatically lower cost than traditional server hosting. DIY only relies on the security of container isolation and a key manager as opposed to the large codebase of a high-level application such as Gmail (and all the Google teams using Gmail data). With attestation technology such as SGX, DIY's execution could also be verified remotely. We show that a DIY email server that sends 500 messages/day costs $0.26/month, which is 50x cheaper than a highly available EC2 server. We also implement a DIY chat service and show that it performs well. Finally, we argue that DIY applications are simple enough to operate that cloud providers could offer a simple "app store" for using them.
今天的网络用户依赖于集中的服务来处理电子邮件、文件传输和聊天等应用程序。不幸的是,这些服务带来了重大的隐私风险:即使有一个善意的提供商,一次违规就可能使数百万用户的数据处于危险之中。另一种选择是让用户托管自己的服务器,但这对大多数应用程序来说都是非常昂贵的:以高可用性模式部署的单个VM每月可能要花费许多美元。在本文中,我们提出了自己部署(DIY),这是一种基于无服务器计算平台(如Amazon Lambda)托管应用程序的新模型。DIY允许用户运行高可用性的服务,比当前的集中式提供商具有更强的隐私保证,并且比传统的服务器托管成本低得多。DIY只依赖于容器隔离和密钥管理器的安全性,而不是像Gmail这样的高级应用程序的大型代码库(以及所有使用Gmail数据的Google团队)。通过SGX等认证技术,DIY的执行也可以远程验证。我们展示了一个每天发送500条消息的DIY电子邮件服务器每月花费0.26美元,比一个高可用性的EC2服务器便宜50倍。我们还实现了一个DIY聊天服务,并证明了它的良好性能。最后,我们认为DIY应用程序足够简单,云提供商可以提供一个简单的“应用程序商店”来使用它们。
{"title":"DIY Hosting for Online Privacy","authors":"Shoumik Palkar, M. Zaharia","doi":"10.1145/3152434.3152459","DOIUrl":"https://doi.org/10.1145/3152434.3152459","url":null,"abstract":"Web users today rely on centralized services for applications such as email, file transfer and chat. Unfortunately, these services create a significant privacy risk: even with a benevolent provider, a single breach can put millions of users' data at risk. One alternative would be for users to host their own servers, but this would be highly expensive for most applications: a single VM deployed in a high-availability mode can cost many dollars per month. In this paper, we propose Deploy It Yourself (DIY), a new model for hosting applications based on serverless computing platforms such as Amazon Lambda. DIY allows users to run a highly available service with much stronger privacy guarantees than current centralized providers, and at a dramatically lower cost than traditional server hosting. DIY only relies on the security of container isolation and a key manager as opposed to the large codebase of a high-level application such as Gmail (and all the Google teams using Gmail data). With attestation technology such as SGX, DIY's execution could also be verified remotely. We show that a DIY email server that sends 500 messages/day costs $0.26/month, which is 50x cheaper than a highly available EC2 server. We also implement a DIY chat service and show that it performs well. Finally, we argue that DIY applications are simple enough to operate that cloud providers could offer a simple \"app store\" for using them.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114259884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
HotCocoa: Hardware Congestion Control Abstractions HotCocoa:硬件拥塞控制抽象
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152457
Mina Tahmasbi Arashloo, Monia Ghobadi, J. Rexford, D. Walker
Congestion control in multi-tenant data centers is an active area of research because of its significant impact on customer experience, and, consequently, on revenue. Therefore, new algorithms and protocols are expected to emerge as the Cloud evolves. Deploying new congestion control algorithms in the end host's hypervisor allows frequent updates, but processing packets at high rates in the hypervisor and implementing the elements of a congestion control algorithm, such as traffic shapers and timestamps, in software have well-studied inaccuracies and CPU inefficiencies. In this paper, we argue for implementing the entire congestion control algorithm in programmable NICs. To do so, we identify the absence of hardware-aware programming abstractions as the most immediate challenge and solve it using a simple high-level domain specific language called HotCocoa. HotCocoa lies at a sweet spot between the ability to express a broad set of congestion control algorithms and efficient hardware implementation. It offers a set of hardware-aware COngestion COntrol Abstractions that enable operators to specify their algorithm without having to worry about low-level hardware primitives. To evaluate HotCocoa, we implement four congestion control algorithms (Reno, DCTCP, PCC, and TIMELY) and use simulations to show that HotCocoa's implementation of Reno perfectly tracks the behavior of a native implementation in C++.
多租户数据中心中的拥塞控制是一个活跃的研究领域,因为它会对客户体验产生重大影响,从而影响收入。因此,随着云的发展,新的算法和协议有望出现。在终端主机的虚拟机管理程序中部署新的拥塞控制算法允许频繁更新,但是在虚拟机管理程序中以高速率处理数据包,并在软件中实现拥塞控制算法的元素,如流量形状器和时间戳,已经得到了充分研究的不准确性和CPU效率低下。在本文中,我们讨论了在可编程网卡中实现整个拥塞控制算法。为此,我们认为缺乏硬件感知编程抽象是最直接的挑战,并使用一种简单的高级领域特定语言HotCocoa来解决这个问题。HotCocoa处于表达广泛的拥塞控制算法和高效硬件实现之间的最佳位置。它提供了一组硬件感知的拥塞控制抽象,使运营商能够指定他们的算法,而不必担心低级硬件原语。为了评估HotCocoa,我们实现了四种拥塞控制算法(Reno, DCTCP, PCC和TIMELY),并使用模拟来显示HotCocoa的Reno实现完美地跟踪了c++中本地实现的行为。
{"title":"HotCocoa: Hardware Congestion Control Abstractions","authors":"Mina Tahmasbi Arashloo, Monia Ghobadi, J. Rexford, D. Walker","doi":"10.1145/3152434.3152457","DOIUrl":"https://doi.org/10.1145/3152434.3152457","url":null,"abstract":"Congestion control in multi-tenant data centers is an active area of research because of its significant impact on customer experience, and, consequently, on revenue. Therefore, new algorithms and protocols are expected to emerge as the Cloud evolves. Deploying new congestion control algorithms in the end host's hypervisor allows frequent updates, but processing packets at high rates in the hypervisor and implementing the elements of a congestion control algorithm, such as traffic shapers and timestamps, in software have well-studied inaccuracies and CPU inefficiencies. In this paper, we argue for implementing the entire congestion control algorithm in programmable NICs. To do so, we identify the absence of hardware-aware programming abstractions as the most immediate challenge and solve it using a simple high-level domain specific language called HotCocoa. HotCocoa lies at a sweet spot between the ability to express a broad set of congestion control algorithms and efficient hardware implementation. It offers a set of hardware-aware COngestion COntrol Abstractions that enable operators to specify their algorithm without having to worry about low-level hardware primitives. To evaluate HotCocoa, we implement four congestion control algorithms (Reno, DCTCP, PCC, and TIMELY) and use simulations to show that HotCocoa's implementation of Reno perfectly tracks the behavior of a native implementation in C++.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"156 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114015563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Stick a fork in it: Analyzing the Ethereum network partition 插入一个分叉:分析以太坊网络分区
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152449
Lucianna Kiffer, Dave Levin, A. Mislove
As blockchain technologies and cryptocurrencies increase in popularity, their decentralization poses unique challenges in network partitions. In traditional distributed systems, network partitions are generally a result of bugs or connectivity failures; the typical goal of the system designer is to automatically recover from such issues as seamlessly as possible. Blockchain-based systems, however, rely on purposeful "forks" to roll out protocol changes in a decentralized manner. Not all users may agree with proposed changes, and thus forks can persist, leading to permanent network partitions. In this paper, we closely study the large-scale fork that occurred in Ethereum, a new blockchain technology that allows for both currency transactions and smart contracts. Ethereum is currently the second-most-valuable cryptocurrency, with a market capitalization of over $28B. We explore the consequences of this fork, showing the impact on the two networks and their mining pools, and how the fork lead to unintentional incentives and security vulnerabilities.
随着区块链技术和加密货币的普及,它们的去中心化给网络分区带来了独特的挑战。在传统的分布式系统中,网络分区通常是错误或连接失败的结果;系统设计人员的典型目标是尽可能无缝地从这些问题中自动恢复。然而,基于区块链的系统依赖于有目的的“分叉”,以分散的方式推出协议更改。并非所有用户都同意提议的更改,因此分叉可能会持续存在,从而导致永久的网络分区。在本文中,我们仔细研究了以太坊中发生的大规模分叉,以太坊是一种新的区块链技术,允许货币交易和智能合约。以太坊目前是第二大最有价值的加密货币,市值超过280亿美元。我们探讨了这次分叉的后果,展示了对两个网络及其矿池的影响,以及分叉如何导致无意的激励和安全漏洞。
{"title":"Stick a fork in it: Analyzing the Ethereum network partition","authors":"Lucianna Kiffer, Dave Levin, A. Mislove","doi":"10.1145/3152434.3152449","DOIUrl":"https://doi.org/10.1145/3152434.3152449","url":null,"abstract":"As blockchain technologies and cryptocurrencies increase in popularity, their decentralization poses unique challenges in network partitions. In traditional distributed systems, network partitions are generally a result of bugs or connectivity failures; the typical goal of the system designer is to automatically recover from such issues as seamlessly as possible. Blockchain-based systems, however, rely on purposeful \"forks\" to roll out protocol changes in a decentralized manner. Not all users may agree with proposed changes, and thus forks can persist, leading to permanent network partitions. In this paper, we closely study the large-scale fork that occurred in Ethereum, a new blockchain technology that allows for both currency transactions and smart contracts. Ethereum is currently the second-most-valuable cryptocurrency, with a market capitalization of over $28B. We explore the consequences of this fork, showing the impact on the two networks and their mining pools, and how the fork lead to unintentional incentives and security vulnerabilities.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121457612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
An Axiomatic Approach to Congestion Control 拥塞控制的公理方法
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152445
Doron Zarchy, R. Mittal, Michael Schapira, S. Shenker
Recent years have witnessed a surge of interest in congestion control. Unfortunately, the overwhelmingly large design space along with the increasingly diverse range of application environments makes evaluating congestion control protocols a daunting task. Researchers often use simulation and experiments to examine the performance of designs in specific contexts, but this gives limited insight into the more general properties of these schemes and provides no information about the inherent limits of congestion control designs, e.g., which properties are simultaneously achievable. To complement simulation and experimentation, we advocate a principled framework for reasoning about congestion control protocols. We report on our initial steps in this direction, which was inspired by the axiomatic approach from social choice theory and game theory. We consider several natural requirements ("axioms") from congestion control protocols -- e.g., efficient resource-utilization, loss-avoidance, fairness, stability, and TCP-friendliness -- and investigate which combinations of these can be achieved within a single design. Thus, our framework allows us to investigate the fundamental tradeoffs between desiderata, and to identify where existing and new congestion control architectures fit within the space of possible outcomes. We believe that our results are but a first step in the axiomatic exploration of congestion control and leave the reader with exciting directions for future research.
近年来,人们对拥堵控制的兴趣激增。不幸的是,巨大的设计空间以及日益多样化的应用程序环境使得评估拥塞控制协议成为一项艰巨的任务。研究人员经常使用模拟和实验来检查特定环境下设计的性能,但这对这些方案的更一般属性提供了有限的见解,并且没有提供关于拥塞控制设计固有限制的信息,例如,哪些属性是同时可实现的。为了补充模拟和实验,我们提倡一个有原则的框架来推理拥塞控制协议。我们报告了我们在这个方向上的初步步骤,这是受到社会选择理论和博弈论的公理化方法的启发。我们考虑了拥塞控制协议的几个自然要求(“公理”)——例如,有效的资源利用、避免损失、公平性、稳定性和tcp友好性——并研究了在单个设计中可以实现这些要求的哪些组合。因此,我们的框架允许我们调查期望之间的基本权衡,并确定现有的和新的拥塞控制架构在可能的结果空间中适合的位置。我们相信,我们的结果只是拥塞控制公理探索的第一步,并为读者留下了令人兴奋的未来研究方向。
{"title":"An Axiomatic Approach to Congestion Control","authors":"Doron Zarchy, R. Mittal, Michael Schapira, S. Shenker","doi":"10.1145/3152434.3152445","DOIUrl":"https://doi.org/10.1145/3152434.3152445","url":null,"abstract":"Recent years have witnessed a surge of interest in congestion control. Unfortunately, the overwhelmingly large design space along with the increasingly diverse range of application environments makes evaluating congestion control protocols a daunting task. Researchers often use simulation and experiments to examine the performance of designs in specific contexts, but this gives limited insight into the more general properties of these schemes and provides no information about the inherent limits of congestion control designs, e.g., which properties are simultaneously achievable. To complement simulation and experimentation, we advocate a principled framework for reasoning about congestion control protocols. We report on our initial steps in this direction, which was inspired by the axiomatic approach from social choice theory and game theory. We consider several natural requirements (\"axioms\") from congestion control protocols -- e.g., efficient resource-utilization, loss-avoidance, fairness, stability, and TCP-friendliness -- and investigate which combinations of these can be achieved within a single design. Thus, our framework allows us to investigate the fundamental tradeoffs between desiderata, and to identify where existing and new congestion control architectures fit within the space of possible outcomes. We believe that our results are but a first step in the axiomatic exploration of congestion control and leave the reader with exciting directions for future research.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114402289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Learning to Route 学习路由
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152441
Asaf Valadarsky, Michael Schapira, Dafna Shahaf, Aviv Tamar
Recently, much attention has been devoted to the question of whether/when traditional network protocol design, which relies on the application of algorithmic insights by human experts, can be replaced by a data-driven (i.e., machine learning) approach. We explore this question in the context of the arguably most fundamental networking task: routing. Can ideas and techniques from machine learning (ML) be leveraged to automatically generate "good" routing configurations? We focus on the classical setting of intradomain traffic engineering. We observe that this context poses significant challenges for data-driven protocol design. Our preliminary results regarding the power of data-driven routing suggest that applying ML (specifically, deep reinforcement learning) to this context yields high performance and is a promising direction for further research. We outline a research agenda for ML-guided routing.
最近,人们非常关注依赖于人类专家算法见解应用的传统网络协议设计是否/何时可以被数据驱动(即机器学习)方法所取代的问题。我们在最基本的网络任务:路由的背景下探讨这个问题。机器学习(ML)的思想和技术能否被用来自动生成“好的”路由配置?我们关注域内流量工程的经典设置。我们观察到,这种情况对数据驱动的协议设计提出了重大挑战。我们关于数据驱动路由能力的初步结果表明,将ML(特别是深度强化学习)应用于这种情况可以产生高性能,并且是进一步研究的有希望的方向。我们概述了机器学习引导路由的研究议程。
{"title":"Learning to Route","authors":"Asaf Valadarsky, Michael Schapira, Dafna Shahaf, Aviv Tamar","doi":"10.1145/3152434.3152441","DOIUrl":"https://doi.org/10.1145/3152434.3152441","url":null,"abstract":"Recently, much attention has been devoted to the question of whether/when traditional network protocol design, which relies on the application of algorithmic insights by human experts, can be replaced by a data-driven (i.e., machine learning) approach. We explore this question in the context of the arguably most fundamental networking task: routing. Can ideas and techniques from machine learning (ML) be leveraged to automatically generate \"good\" routing configurations? We focus on the classical setting of intradomain traffic engineering. We observe that this context poses significant challenges for data-driven protocol design. Our preliminary results regarding the power of data-driven routing suggest that applying ML (specifically, deep reinforcement learning) to this context yields high performance and is a promising direction for further research. We outline a research agenda for ML-guided routing.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128459057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 156
Biases in Data-Driven Networking, and What to Do About Them 数据驱动网络中的偏见,以及如何解决它们
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152448
Mihovil Bartulovic, Junchen Jiang, Sivaraman Balakrishnan, V. Sekar, B. Sinopoli
Recent efforts highlight the promise of data-driven approaches to optimize network decisions. Many such efforts use trace-driven evaluation; i.e., running offline analysis on network traces to estimate the potential benefits of different policies before running them in practice. Unfortunately, such frameworks can have fundamental pitfalls (e.g., skews due to previous policies that were used in the data collection phase and insufficient data for specific subpopulations) that could lead to misleading estimates and ultimately suboptimal decisions. In this paper, we shed light on such pitfalls and identify a promising roadmap to address these pitfalls by leveraging parallels in causal inference, namely the Doubly Robust estimator.
最近的努力突出了数据驱动方法优化网络决策的前景。许多这样的工作使用跟踪驱动的评估;即,在实际运行不同策略之前,对网络轨迹进行离线分析,以估计不同策略的潜在收益。不幸的是,这种框架可能存在根本性的缺陷(例如,由于以前在数据收集阶段使用的政策和特定亚群体的数据不足而产生的偏差),这可能导致误导性估计并最终导致次优决策。在本文中,我们阐明了这样的陷阱,并确定了一个有前途的路线图,通过利用因果推理中的相似之处来解决这些陷阱,即双重鲁棒估计器。
{"title":"Biases in Data-Driven Networking, and What to Do About Them","authors":"Mihovil Bartulovic, Junchen Jiang, Sivaraman Balakrishnan, V. Sekar, B. Sinopoli","doi":"10.1145/3152434.3152448","DOIUrl":"https://doi.org/10.1145/3152434.3152448","url":null,"abstract":"Recent efforts highlight the promise of data-driven approaches to optimize network decisions. Many such efforts use trace-driven evaluation; i.e., running offline analysis on network traces to estimate the potential benefits of different policies before running them in practice. Unfortunately, such frameworks can have fundamental pitfalls (e.g., skews due to previous policies that were used in the data collection phase and insufficient data for specific subpopulations) that could lead to misleading estimates and ultimately suboptimal decisions. In this paper, we shed light on such pitfalls and identify a promising roadmap to address these pitfalls by leveraging parallels in causal inference, namely the Doubly Robust estimator.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122778934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
FreeLab: A Free Experimentation Platform 一个免费的实验平台
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152436
Matteo Varvello, Diego Perino
As researchers, we are aware of how hard it is to obtain access to vantage points in the Internet. Experimentation platforms are useful tools, but they are also: 1) paid, either via a membership fee or by resource sharing, 2) unreliable, nodes come and go, 3) outdated, often still run on their original hardware and OS. While one could build yet-another platform with up-to-date and reliable hardware and software, it is hard to imagine one which is free. This is the goal of this paper: we set out to build FreeLab, a free experimentation platform which also aims to be reliable and up-to-date. The key idea behind FreeLab is that experiments run directly at its user machines, while traffic is relayed by free vantage points in the Internet (web and SOCKS proxies, and DNS resolvers). FreeLab is thus free of access by design and up-to-date as far as its users maintain their experimenting machines. Reliability is a key challenge due to the volatile nature of free resources, and the introduction of errors (path inflation, header manipulation, bandwidth shrinkage) caused by traffic relays.
作为研究人员,我们意识到在互联网上获得有利位置是多么困难。实验平台是有用的工具,但它们也是:1)付费的,要么通过会员费,要么通过资源共享;2)不可靠,节点来来往往;3)过时的,通常仍然在原来的硬件和操作系统上运行。虽然人们可以用最新的、可靠的硬件和软件构建另一个平台,但很难想象一个是免费的。这就是本文的目标:我们着手建立FreeLab,一个免费的实验平台,它的目标也是可靠和最新的。FreeLab背后的关键思想是,实验直接在其用户机器上运行,而流量则通过互联网上的免费有利位置(web和SOCKS代理以及DNS解析器)进行中继。因此,只要用户维护他们的实验机器,FreeLab的设计是免费的,并且是最新的。可靠性是一个关键的挑战,因为空闲资源的不稳定性,以及由流量中继引起的错误(路径膨胀、报头操作、带宽缩减)的引入。
{"title":"FreeLab: A Free Experimentation Platform","authors":"Matteo Varvello, Diego Perino","doi":"10.1145/3152434.3152436","DOIUrl":"https://doi.org/10.1145/3152434.3152436","url":null,"abstract":"As researchers, we are aware of how hard it is to obtain access to vantage points in the Internet. Experimentation platforms are useful tools, but they are also: 1) paid, either via a membership fee or by resource sharing, 2) unreliable, nodes come and go, 3) outdated, often still run on their original hardware and OS. While one could build yet-another platform with up-to-date and reliable hardware and software, it is hard to imagine one which is free. This is the goal of this paper: we set out to build FreeLab, a free experimentation platform which also aims to be reliable and up-to-date. The key idea behind FreeLab is that experiments run directly at its user machines, while traffic is relayed by free vantage points in the Internet (web and SOCKS proxies, and DNS resolvers). FreeLab is thus free of access by design and up-to-date as far as its users maintain their experimenting machines. Reliability is a key challenge due to the volatile nature of free resources, and the introduction of errors (path inflation, header manipulation, bandwidth shrinkage) caused by traffic relays.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129867311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
In-Network Computation is a Dumb Idea Whose Time Has Come 网络计算是一个愚蠢的想法,它的时代已经到来
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152461
Amedeo Sapio, I. Abdelaziz, Abdulla Aldilaijan, M. Canini, Panos Kalnis
Programmable data plane hardware creates new opportunities for infusing intelligence into the network. This raises a fundamental question: what kinds of computation should be delegated to the network? In this paper, we discuss the opportunities and challenges for co-designing data center distributed systems with their network layer. We believe that the time has finally come for offloading part of their computation to execute in-network. However, in-network computation tasks must be judiciously crafted to match the limitations of the network machine architecture of programmable devices. With the help of our experiments on machine learning and graph analytics workloads, we identify that aggregation functions raise opportunities to exploit the limited computation power of networking hardware to lessen network congestion and improve the overall application performance. Moreover, as a proof-of-concept, we propose Daiet, a system that performs in-network data aggregation. Experimental results with an initial prototype show a large data reduction ratio (86.9%-89.3%) and a similar decrease in the workers' computation time.
可编程数据平面硬件为向网络中注入智能创造了新的机会。这就提出了一个基本问题:什么样的计算应该委托给网络?本文讨论了数据中心分布式系统及其网络层协同设计的机遇和挑战。我们相信,将它们的部分计算卸载到网络内执行的时机终于到来了。然而,网络内的计算任务必须谨慎地设计,以匹配可编程设备的网络机器体系结构的限制。借助我们在机器学习和图形分析工作负载上的实验,我们发现聚合功能增加了利用网络硬件有限的计算能力来减少网络拥塞和提高整体应用程序性能的机会。此外,作为概念验证,我们提出了Daiet,一个执行网络内数据聚合的系统。初始样机的实验结果表明,该算法具有较大的数据缩减率(86.9% ~ 89.3%),并且工人的计算时间也有类似的减少。
{"title":"In-Network Computation is a Dumb Idea Whose Time Has Come","authors":"Amedeo Sapio, I. Abdelaziz, Abdulla Aldilaijan, M. Canini, Panos Kalnis","doi":"10.1145/3152434.3152461","DOIUrl":"https://doi.org/10.1145/3152434.3152461","url":null,"abstract":"Programmable data plane hardware creates new opportunities for infusing intelligence into the network. This raises a fundamental question: what kinds of computation should be delegated to the network? In this paper, we discuss the opportunities and challenges for co-designing data center distributed systems with their network layer. We believe that the time has finally come for offloading part of their computation to execute in-network. However, in-network computation tasks must be judiciously crafted to match the limitations of the network machine architecture of programmable devices. With the help of our experiments on machine learning and graph analytics workloads, we identify that aggregation functions raise opportunities to exploit the limited computation power of networking hardware to lessen network congestion and improve the overall application performance. Moreover, as a proof-of-concept, we propose Daiet, a system that performs in-network data aggregation. Experimental results with an initial prototype show a large data reduction ratio (86.9%-89.3%) and a similar decrease in the workers' computation time.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"2013 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129907771","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 203
Harvesting Randomness to Optimize Distributed Systems 收集随机性以优化分布式系统
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152435
Mathias Lécuyer, Joshua Lockerman, Lamont Nelson, S. Sen, Amit Sharma, Aleksandrs Slivkins
We view randomization through the lens of statistical machine learning: as a powerful resource for offline optimization. Cloud systems make randomized decisions all the time (e.g., in load balancing), yet this randomness is rarely used for optimization after-the-fact. By casting system decisions in the framework of reinforcement learning, we show how to collect data from existing systems, without modifying them, to evaluate new policies, without deploying them. Our methodology, called harvesting randomness, has the potential to accurately estimate a policy's performance without the risk or cost of deploying it on live traffic. We quantify this optimization power and apply it to a real machine health scenario in Azure Compute. We also apply it to two prototyped scenarios, for load balancing (Nginx) and caching (Redis), with much less success, and use them to identify the systems and machine learning challenges to achieving our goal. Our long-term agenda is to harvest the randomness in distributed systems to develop non-invasive and efficient techniques for optimizing them. Like CPU cycles and bandwidth, we view randomness as a valuable resource being wasted by the cloud, and we seek to remedy this.
我们通过统计机器学习的视角来看待随机化:作为离线优化的强大资源。云系统总是做出随机决策(例如,在负载平衡中),但这种随机性很少用于事后优化。通过在强化学习框架中进行系统决策,我们展示了如何在不修改现有系统的情况下从现有系统中收集数据,以评估新策略,而无需部署它们。我们的方法,称为收集随机性,有可能准确地估计策略的性能,而不会在实时流量上部署它的风险或成本。我们量化了这种优化能力,并将其应用于Azure Compute中的真实机器运行状况场景。我们还将其应用于两个原型场景,用于负载平衡(Nginx)和缓存(Redis),但成功率低得多,并使用它们来识别系统和机器学习挑战,以实现我们的目标。我们的长期议程是收获分布式系统中的随机性,以开发非侵入性和有效的技术来优化它们。就像CPU周期和带宽一样,我们将随机性视为被云浪费的宝贵资源,并寻求补救。
{"title":"Harvesting Randomness to Optimize Distributed Systems","authors":"Mathias Lécuyer, Joshua Lockerman, Lamont Nelson, S. Sen, Amit Sharma, Aleksandrs Slivkins","doi":"10.1145/3152434.3152435","DOIUrl":"https://doi.org/10.1145/3152434.3152435","url":null,"abstract":"We view randomization through the lens of statistical machine learning: as a powerful resource for offline optimization. Cloud systems make randomized decisions all the time (e.g., in load balancing), yet this randomness is rarely used for optimization after-the-fact. By casting system decisions in the framework of reinforcement learning, we show how to collect data from existing systems, without modifying them, to evaluate new policies, without deploying them. Our methodology, called harvesting randomness, has the potential to accurately estimate a policy's performance without the risk or cost of deploying it on live traffic. We quantify this optimization power and apply it to a real machine health scenario in Azure Compute. We also apply it to two prototyped scenarios, for load balancing (Nginx) and caching (Redis), with much less success, and use them to identify the systems and machine learning challenges to achieving our goal. Our long-term agenda is to harvest the randomness in distributed systems to develop non-invasive and efficient techniques for optimizing them. Like CPU cycles and bandwidth, we view randomness as a valuable resource being wasted by the cloud, and we seek to remedy this.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132518908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Online Advertising under Internet Censorship 互联网审查下的网络广告
Pub Date : 2017-11-30 DOI: 10.1145/3152434.3152455
Hira Javaid, Hafiz Kamran Khalil, Z. A. Uzmi, I. Qazi
Online advertising plays a critical role in enabling the free Web by allowing publishers to monetize their services. However, the rise in internet censorship events globally poses an economic threat to the advertising ecosystem. This paper studies this interplay and presents Advention, a system that provides censorship circumvention while serving relevant ads. Advention leverages the observation that ad systems are usually hosted on domains that are different from the publisher domains and are almost always uncensored. Taking cue from this, Advention fetches ads via the direct, uncensored, channel between users and the ad system. Preliminary results show that Advention not only offers high ad relevance compared to other popular relay-based circumvention tools, it also offers smaller page load times.
在线广告通过允许出版商从他们的服务中获利,在实现免费网络方面发挥着关键作用。然而,全球互联网审查事件的增加对广告生态系统构成了经济威胁。本文研究了这种相互作用,并提出了Advention,这是一个在提供相关广告的同时提供审查规避的系统。Advention利用了广告系统通常托管在不同于出版商域的域上,并且几乎总是未经审查的观察结果。受此启发,Advention通过用户和广告系统之间的直接、未经审查的渠道获取广告。初步结果表明,与其他流行的基于中继的翻墙工具相比,Advention不仅提供了高广告相关性,而且还提供了更短的页面加载时间。
{"title":"Online Advertising under Internet Censorship","authors":"Hira Javaid, Hafiz Kamran Khalil, Z. A. Uzmi, I. Qazi","doi":"10.1145/3152434.3152455","DOIUrl":"https://doi.org/10.1145/3152434.3152455","url":null,"abstract":"Online advertising plays a critical role in enabling the free Web by allowing publishers to monetize their services. However, the rise in internet censorship events globally poses an economic threat to the advertising ecosystem. This paper studies this interplay and presents Advention, a system that provides censorship circumvention while serving relevant ads. Advention leverages the observation that ad systems are usually hosted on domains that are different from the publisher domains and are almost always uncensored. Taking cue from this, Advention fetches ads via the direct, uncensored, channel between users and the ad system. Preliminary results show that Advention not only offers high ad relevance compared to other popular relay-based circumvention tools, it also offers smaller page load times.","PeriodicalId":120886,"journal":{"name":"Proceedings of the 16th ACM Workshop on Hot Topics in Networks","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114970687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Proceedings of the 16th ACM Workshop on Hot Topics in Networks
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1