首页 > 最新文献

Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing最新文献

英文 中文
Data-dependent hashing via nonlinear spectral gaps 基于非线性谱隙的数据相关哈希
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188846
Alexandr Andoni, A. Naor, Aleksandar Nikolov, Ilya P. Razenshteyn, Erik Waingarten
We establish a generic reduction from _nonlinear spectral gaps_ of metric spaces to data-dependent Locality-Sensitive Hashing, yielding a new approach to the high-dimensional Approximate Near Neighbor Search problem (ANN) under various distance functions. Using this reduction, we obtain the following results: * For _general_ d-dimensional normed spaces and n-point datasets, we obtain a _cell-probe_ ANN data structure with approximation O(logd/ε2), space dO(1) n1+ε, and dO(1)nε cell probes per query, for any ε>0. No non-trivial approximation was known before in this generality other than the O(√d) bound which follows from embedding a general norm into ℓ2. * For ℓp and Schatten-p norms, we improve the data structure further, to obtain approximation O(p) and sublinear query _time_. For ℓp, this improves upon the previous best approximation 2O(p) (which required polynomial as opposed to near-linear in n space). For the Schatten-p norm, no non-trivial ANN data structure was known before this work. Previous approaches to the ANN problem either exploit the low dimensionality of a metric, requiring space exponential in the dimension, or circumvent the curse of dimensionality by embedding a metric into a ”tractable” space, such as ℓ1. Our new generic reduction proceeds differently from both of these approaches using a novel partitioning method.
我们建立了从度量空间的非线性谱隙到数据相关的位置敏感哈希的一般约简,给出了一种解决各种距离函数下的高维近似近邻搜索问题的新方法。*对于一般d维归范空间和n点数据集,我们得到了近似为O(logd/ε2)的_cell-probe_ ANN数据结构,对于任意ε>,每次查询的空间dO(1) n1+ε和dO(1)nε细胞探针。在此一般性中,除了将一般范数嵌入到l2中所得到的O(√d)界之外,没有已知的非平凡近似。*对于p和schattenp范数,我们进一步改进了数据结构,得到近似O(p)和次线性查询_time_。对于p,这改进了之前的最佳近似2O(p)(它需要多项式,而不是n空间中的近线性)。对于schattenp范数,在此工作之前没有已知的非平凡ANN数据结构。以前的人工神经网络问题的方法要么利用度量的低维,在维度上需要空间指数,要么通过将度量嵌入到“可处理”的空间(如1)来规避维度的诅咒。我们的新通用约简使用一种新的划分方法,与这两种方法不同。
{"title":"Data-dependent hashing via nonlinear spectral gaps","authors":"Alexandr Andoni, A. Naor, Aleksandar Nikolov, Ilya P. Razenshteyn, Erik Waingarten","doi":"10.1145/3188745.3188846","DOIUrl":"https://doi.org/10.1145/3188745.3188846","url":null,"abstract":"We establish a generic reduction from _nonlinear spectral gaps_ of metric spaces to data-dependent Locality-Sensitive Hashing, yielding a new approach to the high-dimensional Approximate Near Neighbor Search problem (ANN) under various distance functions. Using this reduction, we obtain the following results: * For _general_ d-dimensional normed spaces and n-point datasets, we obtain a _cell-probe_ ANN data structure with approximation O(logd/ε2), space dO(1) n1+ε, and dO(1)nε cell probes per query, for any ε>0. No non-trivial approximation was known before in this generality other than the O(√d) bound which follows from embedding a general norm into ℓ2. * For ℓp and Schatten-p norms, we improve the data structure further, to obtain approximation O(p) and sublinear query _time_. For ℓp, this improves upon the previous best approximation 2O(p) (which required polynomial as opposed to near-linear in n space). For the Schatten-p norm, no non-trivial ANN data structure was known before this work. Previous approaches to the ANN problem either exploit the low dimensionality of a metric, requiring space exponential in the dimension, or circumvent the curse of dimensionality by embedding a metric into a ”tractable” space, such as ℓ1. Our new generic reduction proceeds differently from both of these approaches using a novel partitioning method.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74885773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Interactive compression to external information 对外部信息的交互式压缩
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188956
M. Braverman, Gillat Kol
We describe a new way of compressing two-party communication protocols to get protocols with potentially smaller communication. We show that every communication protocol that communicates C bits and reveals I bits of information about the participants’ private inputs to an observer that watches the communication, can be simulated by a new protocol that communicates at most poly(I) · loglog(C) bits. Our result is tight up to polynomial factors, as it matches the recent work separating communication complexity from external information cost.
我们描述了一种压缩双方通信协议的新方法,以获得具有潜在更小通信的协议。我们表明,每个通信协议通信C位,并向观察通信的观察者显示有关参与者私有输入的I位信息,可以通过最多通信poly(I)·loglog(C)位的新协议来模拟。我们的结果与多项式因子紧密相关,因为它与最近将通信复杂性与外部信息成本分离的工作相匹配。
{"title":"Interactive compression to external information","authors":"M. Braverman, Gillat Kol","doi":"10.1145/3188745.3188956","DOIUrl":"https://doi.org/10.1145/3188745.3188956","url":null,"abstract":"We describe a new way of compressing two-party communication protocols to get protocols with potentially smaller communication. We show that every communication protocol that communicates C bits and reveals I bits of information about the participants’ private inputs to an observer that watches the communication, can be simulated by a new protocol that communicates at most poly(I) · loglog(C) bits. Our result is tight up to polynomial factors, as it matches the recent work separating communication complexity from external information cost.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75483300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Hitting sets with near-optimal error for read-once branching programs 对于只读一次的分支程序,命中集的错误接近最优
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188780
M. Braverman, Gil Cohen, Sumegha Garg
Nisan (Combinatorica’92) constructed a pseudorandom generator for length n, width n read-once branching programs (ROBPs) with error ε and seed length O(log2n + logn · log(1/ε)). A major goal in complexity theory is to reduce the seed length, hopefully, to the optimal O(logn+log(1/ε)), or to construct improved hitting sets, as these would yield stronger derandomization of BPL and RL, respectively. In contrast to a successful line of work in restricted settings, no progress has been made for general, unrestricted, ROBPs. Indeed, Nisan’s construction is the best pseudorandom generator and, prior to this work, also the best hitting set for unrestricted ROBPs. In this work, we make the first improvement for the general case by constructing a hitting set with seed length O(log2n+log(1/ε)). That is, we decouple ε and n, and obtain near-optimal dependence on the former. The regime of parameters in which our construction strictly improves upon prior works, namely, log(1/ε) ≫ logn, is well-motivated by the work of Saks and Zhou (J.CSS’99) who use pseudorandom generators with error ε = 2−(logn)2 in their proof for BPL ⊆ L3/2. We further suggest a research program towards proving that BPL ⊆ L4/3 in which our result achieves one step. As our main technical tool, we introduce and construct a new type of primitive we call pseudorandom pseudo-distributions. Informally, this is a generalization of pseudorandom generators in which one may assign negative and unbounded weights to paths as opposed to working with probability distributions. We show that such a primitive yields hitting sets and, for derandomization purposes, can be used to derandomize two-sided error algorithms.
Nisan (Combinatorica ' 92)构建了一个伪随机生成器,用于长度为n,宽度为n,误差为ε,种子长度为O(log2n + logn·log(1/ε))的只读一次分支程序(robp)。复杂性理论的一个主要目标是将种子长度减少到最优的O(logn+log(1/ε)),或者构建改进的命中集,因为这些将分别产生更强的BPL和RL的非随机化。与限制环境下的成功工作相比,一般的、不受限制的robp没有取得任何进展。事实上,Nisan的构造是最好的伪随机生成器,在此工作之前,也是无限制robp的最佳命中集。在这项工作中,我们通过构造一个种子长度为O(log2n+log(1/ε))的命中集,对一般情况进行了第一次改进。也就是说,我们解耦了ε和n,并获得了对前者的近似最优依赖。Saks和Zhou (J.CSS ' 99)使用误差为ε = 2−(logn)2的伪随机生成器对BPL≥3/2的证明,很好地推动了我们的构造严格改进于先前工作的参数体系,即log(1/ε) > logn) > logn。我们进一步提出了一个研究方案,以证明我们的结果实现了一步。作为我们的主要技术工具,我们引入并构造了一种新的基元,我们称之为伪随机伪分布。非正式地说,这是伪随机生成器的泛化,其中可以为路径分配负的无界权重,而不是使用概率分布。我们证明了这样的原语产生命中集,并且对于非随机化的目的,可以用于非随机化双边误差算法。
{"title":"Hitting sets with near-optimal error for read-once branching programs","authors":"M. Braverman, Gil Cohen, Sumegha Garg","doi":"10.1145/3188745.3188780","DOIUrl":"https://doi.org/10.1145/3188745.3188780","url":null,"abstract":"Nisan (Combinatorica’92) constructed a pseudorandom generator for length n, width n read-once branching programs (ROBPs) with error ε and seed length O(log2n + logn · log(1/ε)). A major goal in complexity theory is to reduce the seed length, hopefully, to the optimal O(logn+log(1/ε)), or to construct improved hitting sets, as these would yield stronger derandomization of BPL and RL, respectively. In contrast to a successful line of work in restricted settings, no progress has been made for general, unrestricted, ROBPs. Indeed, Nisan’s construction is the best pseudorandom generator and, prior to this work, also the best hitting set for unrestricted ROBPs. In this work, we make the first improvement for the general case by constructing a hitting set with seed length O(log2n+log(1/ε)). That is, we decouple ε and n, and obtain near-optimal dependence on the former. The regime of parameters in which our construction strictly improves upon prior works, namely, log(1/ε) ≫ logn, is well-motivated by the work of Saks and Zhou (J.CSS’99) who use pseudorandom generators with error ε = 2−(logn)2 in their proof for BPL ⊆ L3/2. We further suggest a research program towards proving that BPL ⊆ L4/3 in which our result achieves one step. As our main technical tool, we introduce and construct a new type of primitive we call pseudorandom pseudo-distributions. Informally, this is a generalization of pseudorandom generators in which one may assign negative and unbounded weights to paths as opposed to working with probability distributions. We show that such a primitive yields hitting sets and, for derandomization purposes, can be used to derandomize two-sided error algorithms.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85211706","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
The minimum Euclidean-norm point in a convex polytope: Wolfe's combinatorial algorithm is exponential 凸多边形的最小欧几里德范数点:Wolfe的组合算法是指数型的
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188820
J. D. Loera, Jamie Haddock, Luis Rademacher
The complexity of Philip Wolfe’s method for the minimum Euclidean-norm point problem over a convex polytope has remained unknown since he proposed the method in 1974. We present the first example that Wolfe’s method takes exponential time. Additionally, we improve previous results to show that linear programming reduces in strongly-polynomial time to the minimum norm point problem over a simplex
Philip Wolfe的凸多面体上最小欧几里得范数点问题的方法自1974年提出以来,其复杂性一直是未知的。我们给出了Wolfe方法需要指数时间的第一个例子。此外,我们改进了先前的结果,表明线性规划在强多项式时间内减少了单纯形上的最小范数点问题
{"title":"The minimum Euclidean-norm point in a convex polytope: Wolfe's combinatorial algorithm is exponential","authors":"J. D. Loera, Jamie Haddock, Luis Rademacher","doi":"10.1145/3188745.3188820","DOIUrl":"https://doi.org/10.1145/3188745.3188820","url":null,"abstract":"The complexity of Philip Wolfe’s method for the minimum Euclidean-norm point problem over a convex polytope has remained unknown since he proposed the method in 1974. We present the first example that Wolfe’s method takes exponential time. Additionally, we improve previous results to show that linear programming reduces in strongly-polynomial time to the minimum norm point problem over a simplex","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80323550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
A tighter welfare guarantee for first-price auctions 为首价拍卖提供更严格的福利保障
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188944
D. Hoy, Sam Taggart, Zihe Wang
This paper proves that the welfare of the first price auction in Bayes-Nash equilibrium is at least a .743-fraction of the welfare of the optimal mechanism assuming agents’ values are independently distributed. The previous best bound was 1−1/e≈.63, derived using smoothness, the standard technique for reasoning about welfare of games in equilibrium. In the worst known example, the first price auction achieves a ≈.869-fraction of the optimal welfare, far better than the theoretical guarantee. Despite this large gap, it was unclear whether the 1−1/e bound was tight. We prove that it is not. Our analysis eschews smoothness, and instead uses the independence assumption on agents’ value distributions to give a more careful accounting of the welfare contribution of agents who win despite not having the highest value.
本文证明了在贝叶斯-纳什均衡中,假设代理的价值是独立分布的,第一次价格拍卖的福利至少是最优机制福利的0.743。之前的最佳界为1−1/e≈。63,使用平滑性推导出来的,平滑性是推理均衡博弈福利的标准技术。在已知最坏的例子中,第一次价格拍卖达到a≈。最优福利的869分之一,远好于理论保证。尽管有这么大的差距,但尚不清楚1−1/e界限是否严密。我们证明它不是。我们的分析避免了平滑性,而是使用了对代理人价值分布的独立性假设,以更仔细地计算那些尽管没有最高价值但获胜的代理人的福利贡献。
{"title":"A tighter welfare guarantee for first-price auctions","authors":"D. Hoy, Sam Taggart, Zihe Wang","doi":"10.1145/3188745.3188944","DOIUrl":"https://doi.org/10.1145/3188745.3188944","url":null,"abstract":"This paper proves that the welfare of the first price auction in Bayes-Nash equilibrium is at least a .743-fraction of the welfare of the optimal mechanism assuming agents’ values are independently distributed. The previous best bound was 1−1/e≈.63, derived using smoothness, the standard technique for reasoning about welfare of games in equilibrium. In the worst known example, the first price auction achieves a ≈.869-fraction of the optimal welfare, far better than the theoretical guarantee. Despite this large gap, it was unclear whether the 1−1/e bound was tight. We prove that it is not. Our analysis eschews smoothness, and instead uses the independence assumption on agents’ value distributions to give a more careful accounting of the welfare contribution of agents who win despite not having the highest value.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81440185","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Shape of diffusion and size of monochromatic region of a two-dimensional spin system 二维自旋系统的扩散形状和单色区大小
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188836
H. Omidvar, M. Franceschetti
We consider an agent-based distributed algorithm with exponentially distributed waiting times in which agents with binary states interact locally over a geometric graph, and based on this interaction and on the value of a common intolerance threshold τ, decide whether to change their states. This model is equivalent to an Asynchronous Cellular Automaton (ACA) with extended Moore neighborhoods, a zero-temperature Ising model with Glauber dynamics, or a Schelling model of self-organized segregation in an open system, and has applications in the analysis of social and biological networks, and spin glasses systems. We prove a shape theorem for the spread of the “affected” nodes during the process dynamics and show that in the steady state, for τ ∈ (τ*,1−τ*) ∖ {1/2}, where τ* ≈ 0.488, the size of the “mono-chromatic region” at the end of the process is at least exponential in the size of the local neighborhood of interaction with probability approaching one as N grows. Combined with previous results on the expected size of the monochromatic region that provide a matching upper bound, this implies that in the steady state the size of the monochromatic region of any agent is exponential with high probability for the mentioned interval of τ. The shape theorem is based on a novel concentration inequality for the spreading time, and provides a precise geometrical description of the process dynamics. The result on the size of the monochromatic region considerably extends our understanding of the steady state. Showing convergence with high probability, it rules out the possibility that only a small fraction of the nodes are eventually contained in large monochromatic regions, which was left open by previous works.
我们考虑了一种基于智能体的分布式算法,该算法具有指数分布的等待时间,其中具有二进制状态的智能体在几何图上局部相互作用,并基于这种相互作用和常见的不容忍阈值τ的值来决定是否改变它们的状态。该模型等效于具有扩展摩尔邻域的异步元胞自动机(ACA)、具有Glauber动力学的零温度Ising模型或开放系统中自组织分离的Schelling模型,并在社会和生物网络以及自旋玻璃系统的分析中具有应用价值。我们证明了过程动力学中“受影响”节点扩展的一个形状定理,并证明了在稳态下,对于τ∈(τ*,1−τ*)∈{1/2},其中τ*≈0.488,过程结束时的“单色区域”的大小与相互作用局部邻域的大小至少呈指数级增长,且随着N的增长,其概率接近于1。结合先前关于提供匹配上界的单色区域的期望大小的结果,这意味着在稳态下,任何代理的单色区域的大小在上述τ区间内以高概率呈指数型。形状定理基于一种新的扩散时间集中不等式,并提供了过程动力学的精确几何描述。关于单色区大小的结果大大扩展了我们对稳态的理解。它显示了高概率的收敛性,排除了只有一小部分节点最终被包含在大的单色区域的可能性,这是以前的工作留下的开放性。
{"title":"Shape of diffusion and size of monochromatic region of a two-dimensional spin system","authors":"H. Omidvar, M. Franceschetti","doi":"10.1145/3188745.3188836","DOIUrl":"https://doi.org/10.1145/3188745.3188836","url":null,"abstract":"We consider an agent-based distributed algorithm with exponentially distributed waiting times in which agents with binary states interact locally over a geometric graph, and based on this interaction and on the value of a common intolerance threshold τ, decide whether to change their states. This model is equivalent to an Asynchronous Cellular Automaton (ACA) with extended Moore neighborhoods, a zero-temperature Ising model with Glauber dynamics, or a Schelling model of self-organized segregation in an open system, and has applications in the analysis of social and biological networks, and spin glasses systems. We prove a shape theorem for the spread of the “affected” nodes during the process dynamics and show that in the steady state, for τ ∈ (τ*,1−τ*) ∖ {1/2}, where τ* ≈ 0.488, the size of the “mono-chromatic region” at the end of the process is at least exponential in the size of the local neighborhood of interaction with probability approaching one as N grows. Combined with previous results on the expected size of the monochromatic region that provide a matching upper bound, this implies that in the steady state the size of the monochromatic region of any agent is exponential with high probability for the mentioned interval of τ. The shape theorem is based on a novel concentration inequality for the spreading time, and provides a precise geometrical description of the process dynamics. The result on the size of the monochromatic region considerably extends our understanding of the steady state. Showing convergence with high probability, it rules out the possibility that only a small fraction of the nodes are eventually contained in large monochromatic regions, which was left open by previous works.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87060415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
The adaptive complexity of maximizing a submodular function 最大化子模函数的自适应复杂度
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188752
Eric Balkanski, Yaron Singer
In this paper we study the adaptive complexity of submodular optimization. Informally, the adaptive complexity of a problem is the minimal number of sequential rounds required to achieve a constant factor approximation when polynomially-many queries can be executed in parallel at each round. Adaptivity is a fundamental concept that is heavily studied in computer science, largely due to the need for parallelizing computation. Somewhat surprisingly, very little is known about adaptivity in submodular optimization. For the canonical problem of maximizing a monotone submodular function under a cardinality constraint, to the best of our knowledge, all that is known to date is that the adaptive complexity is between 1 and Ω(n). Our main result in this paper is a tight characterization showing that the adaptive complexity of maximizing a monotone submodular function under a cardinality constraint is Θ(log n): - We describe an algorithm which requires O(log n) sequential rounds and achieves an approximation that is arbitrarily close to 1/3; - We show that no algorithm can achieve an approximation better than O(1 / log n) with fewer than O(log n / log log n) rounds. Thus, when allowing for parallelization, our algorithm achieves a constant factor approximation exponentially faster than any known existing algorithm for submodular maximization. Importantly, the approximation algorithm is achieved via adaptive sampling and complements a recent line of work on optimization of functions learned from data. In many cases we do not know the functions we optimize and learn them from labeled samples. Recent results show that no algorithm can obtain a constant factor approximation guarantee using polynomially-many labeled samples as in the PAC and PMAC models, drawn from any distribution. Since learning with non-adaptive samples over any distribution results in a sharp impossibility, we consider learning with adaptive samples where the learner obtains poly(n) samples drawn from a distribution of her choice in every round. Our result implies that in the realizable case, where there is a true underlying function generating the data, Θ(log n) batches of adaptive samples are necessary and sufficient to approximately “learn to optimize” a monotone submodular function under a cardinality constraint.
本文研究了子模优化的自适应复杂度问题。非正式地说,问题的自适应复杂性是在每轮可以并行执行多个多项式查询时,实现常数因子近似值所需的最小连续轮数。自适应是计算机科学中被大量研究的一个基本概念,主要是由于并行计算的需要。令人惊讶的是,我们对子模块优化中的自适应性知之甚少。对于在基数约束下最大化单调次模函数的规范问题,据我们所知,迄今为止所知道的是自适应复杂度在1到Ω(n)之间。我们在本文中的主要结果是一个严格的表征,表明在基数约束下最大化单调子模函数的自适应复杂度为Θ(log n): -我们描述了一个算法,它需要O(log n)个连续轮,并实现任意接近1/3的近似值;-我们证明了没有算法可以在少于O(log n / log n)轮的情况下获得比O(1 / log n)更好的近似。因此,当允许并行化时,我们的算法实现常数因子近似的速度比任何已知的现有的次模最大化算法都要快。重要的是,近似算法是通过自适应采样实现的,并补充了最近从数据中学习的函数优化的一系列工作。在许多情况下,我们不知道我们优化的函数,并从标记的样本中学习它们。最近的研究结果表明,没有一种算法可以像PAC和PMAC模型那样,从任何分布中提取多项式多标记样本,从而获得常因子近似保证。由于在任何分布上使用非自适应样本进行学习都会导致明显的不可能性,因此我们考虑使用自适应样本进行学习,其中学习者在每轮中从她选择的分布中获得多(n)个样本。我们的结果表明,在可实现的情况下,有一个真正的底层函数生成数据,Θ(log n)批自适应样本是必要的,足以在基数约束下近似地“学习优化”单调子模函数。
{"title":"The adaptive complexity of maximizing a submodular function","authors":"Eric Balkanski, Yaron Singer","doi":"10.1145/3188745.3188752","DOIUrl":"https://doi.org/10.1145/3188745.3188752","url":null,"abstract":"In this paper we study the adaptive complexity of submodular optimization. Informally, the adaptive complexity of a problem is the minimal number of sequential rounds required to achieve a constant factor approximation when polynomially-many queries can be executed in parallel at each round. Adaptivity is a fundamental concept that is heavily studied in computer science, largely due to the need for parallelizing computation. Somewhat surprisingly, very little is known about adaptivity in submodular optimization. For the canonical problem of maximizing a monotone submodular function under a cardinality constraint, to the best of our knowledge, all that is known to date is that the adaptive complexity is between 1 and Ω(n). Our main result in this paper is a tight characterization showing that the adaptive complexity of maximizing a monotone submodular function under a cardinality constraint is Θ(log n): - We describe an algorithm which requires O(log n) sequential rounds and achieves an approximation that is arbitrarily close to 1/3; - We show that no algorithm can achieve an approximation better than O(1 / log n) with fewer than O(log n / log log n) rounds. Thus, when allowing for parallelization, our algorithm achieves a constant factor approximation exponentially faster than any known existing algorithm for submodular maximization. Importantly, the approximation algorithm is achieved via adaptive sampling and complements a recent line of work on optimization of functions learned from data. In many cases we do not know the functions we optimize and learn them from labeled samples. Recent results show that no algorithm can obtain a constant factor approximation guarantee using polynomially-many labeled samples as in the PAC and PMAC models, drawn from any distribution. Since learning with non-adaptive samples over any distribution results in a sharp impossibility, we consider learning with adaptive samples where the learner obtains poly(n) samples drawn from a distribution of her choice in every round. Our result implies that in the realizable case, where there is a true underlying function generating the data, Θ(log n) batches of adaptive samples are necessary and sufficient to approximately “learn to optimize” a monotone submodular function under a cardinality constraint.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83388415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 110
Distribution-free junta testing 免费分发军政府测试
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188842
Zhengyang Liu, Xi Chen, R. Servedio, Ying Sheng, Jinyu Xie
We study the problem of testing whether an unknown n-variable Boolean function is a k-junta in the distribution-free property testing model, where the distance between functions is measured with respect to an arbitrary and unknown probability distribution over {0,1}n. Our first main result is that distribution-free k-junta testing can be performed, with one-sided error, by an adaptive algorithm that uses Õ(k2)/є queries (independent of n). Complementing this, our second main result is a lower bound showing that any non-adaptive distribution-free k-junta testing algorithm must make Ω(2k/3) queries even to test to accuracy є=1/3. These bounds establish that while the optimal query complexity of non-adaptive k-junta testing is 2Θ(k), for adaptive testing it is poly(k), and thus show that adaptivity provides an exponential improvement in the distribution-free query complexity of testing juntas.
我们研究了在无分布性质检验模型中检验未知n变量布尔函数是否为k-军政府的问题,其中函数之间的距离是相对于{0,1}n上的任意未知概率分布来测量的。我们的第一个主要结果是,可以通过使用Õ(k2)/ n查询(独立于n)的自适应算法执行无分布k-军政府测试,具有单侧误差。与此相补充的是,我们的第二个主要结果是一个下界,表明任何非自适应无分布k-军政府测试算法必须进行Ω(2k/3)查询,即使测试精度为n =1/3。这些边界表明,非自适应k-军政府测试的最优查询复杂度为2Θ(k),而自适应测试的最优查询复杂度为poly(k),从而表明自适应性在测试军政府的无分布查询复杂度方面提供了指数级的改进。
{"title":"Distribution-free junta testing","authors":"Zhengyang Liu, Xi Chen, R. Servedio, Ying Sheng, Jinyu Xie","doi":"10.1145/3188745.3188842","DOIUrl":"https://doi.org/10.1145/3188745.3188842","url":null,"abstract":"We study the problem of testing whether an unknown n-variable Boolean function is a k-junta in the distribution-free property testing model, where the distance between functions is measured with respect to an arbitrary and unknown probability distribution over {0,1}n. Our first main result is that distribution-free k-junta testing can be performed, with one-sided error, by an adaptive algorithm that uses Õ(k2)/є queries (independent of n). Complementing this, our second main result is a lower bound showing that any non-adaptive distribution-free k-junta testing algorithm must make Ω(2k/3) queries even to test to accuracy є=1/3. These bounds establish that while the optimal query complexity of non-adaptive k-junta testing is 2Θ(k), for adaptive testing it is poly(k), and thus show that adaptivity provides an exponential improvement in the distribution-free query complexity of testing juntas.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82025559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Towards a proof of the 2-to-1 games conjecture? 为了证明2比1博弈猜想?
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188804
Irit Dinur, Subhash Khot, Guy Kindler, Dor Minzer, S. Safra
We present a polynomial time reduction from gap-3LIN to label cover with 2-to-1 constraints. In the “yes” case the fraction of satisfied constraints is at least 1 −ε, and in the “no” case we show that this fraction is at most ε, assuming a certain (new) combinatorial hypothesis on the Grassmann graph. In other words, we describe a combinatorial hypothesis that implies the 2-to-1 conjecture with imperfect completeness. The companion submitted paper [Dinur, Khot, Kindler, Minzer and Safra, STOC 2018] makes some progress towards proving this hypothesis. Our work builds on earlier work by a subset of the authors [Khot, Minzer and Safra, STOC 2017] where a slightly different hypothesis was used to obtain hardness of approximating vertex cover to within factor of √2−ε. The most important implication of this work is (assuming the hypothesis) an NP-hardness gap of 1/2−ε vs. ε for unique games. In addition, we derive optimal NP-hardness for approximating the max-cut-gain problem, NP-hardness of coloring an almost 4-colorable graph with any constant number of colors, and the same √2−ε NP-hardness for approximate vertex cover that was already obtained based on a slightly different hypothesis. Recent progress towards proving our hypothesis [Barak, Kothari and Steurer, ECCC TR18-077], [Dinur, Khot, Kindler, Minzer and Safra, STOC 2018] directly implies some new unconditional NP-hardness results. These include new points of NP-hardness for unique games and for 2-to-1 and 2-to-2 games. More recently, the full version of our hypothesis was proven [Khot, Minzer and Safra, ECCC TR18-006].
我们提出了从gap-3LIN到具有2对1约束的标签覆盖的多项式时间缩减。在“是”的情况下,满足约束的分数至少是1−ε,而在“否”的情况下,我们证明这个分数最多是ε,假设在Grassmann图上的某个(新的)组合假设。换句话说,我们描述了一个包含不完全完备性2比1猜想的组合假设。同行提交的论文[Dinur, Khot, Kindler, Minzer和Safra, STOC 2018]在证明这一假设方面取得了一些进展。我们的工作建立在作者子集的早期工作的基础上[Khot, Minzer和Safra, STOC 2017],其中使用了一个略有不同的假设来获得近似顶点覆盖的硬度在√2−ε因子内。这项工作最重要的含义是(假设假设)np -硬度差距为1/2−ε与ε对于唯一游戏。此外,我们还推导出了逼近最大切割增益问题的最优np -硬度,具有任意常数颜色的几乎4色图着色的np -硬度,以及基于稍微不同的假设已经获得的近似顶点覆盖的相同√2−ε np -硬度。最近证明我们假设的进展[Barak, Kothari和Steurer, ECCC TR18-077], [Dinur, Khot, Kindler, Minzer和Safra, STOC 2018]直接暗示了一些新的无条件np -硬度结果。这包括针对独特游戏以及2对1和2对2游戏的新np硬度点。最近,我们的假设的完整版本被证明[Khot, Minzer和Safra, ECCC TR18-006]。
{"title":"Towards a proof of the 2-to-1 games conjecture?","authors":"Irit Dinur, Subhash Khot, Guy Kindler, Dor Minzer, S. Safra","doi":"10.1145/3188745.3188804","DOIUrl":"https://doi.org/10.1145/3188745.3188804","url":null,"abstract":"We present a polynomial time reduction from gap-3LIN to label cover with 2-to-1 constraints. In the “yes” case the fraction of satisfied constraints is at least 1 −ε, and in the “no” case we show that this fraction is at most ε, assuming a certain (new) combinatorial hypothesis on the Grassmann graph. In other words, we describe a combinatorial hypothesis that implies the 2-to-1 conjecture with imperfect completeness. The companion submitted paper [Dinur, Khot, Kindler, Minzer and Safra, STOC 2018] makes some progress towards proving this hypothesis. Our work builds on earlier work by a subset of the authors [Khot, Minzer and Safra, STOC 2017] where a slightly different hypothesis was used to obtain hardness of approximating vertex cover to within factor of √2−ε. The most important implication of this work is (assuming the hypothesis) an NP-hardness gap of 1/2−ε vs. ε for unique games. In addition, we derive optimal NP-hardness for approximating the max-cut-gain problem, NP-hardness of coloring an almost 4-colorable graph with any constant number of colors, and the same √2−ε NP-hardness for approximate vertex cover that was already obtained based on a slightly different hypothesis. Recent progress towards proving our hypothesis [Barak, Kothari and Steurer, ECCC TR18-077], [Dinur, Khot, Kindler, Minzer and Safra, STOC 2018] directly implies some new unconditional NP-hardness results. These include new points of NP-hardness for unique games and for 2-to-1 and 2-to-2 games. More recently, the full version of our hypothesis was proven [Khot, Minzer and Safra, ECCC TR18-006].","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79657507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 75
Fast algorithms for knapsack via convolution and prediction 基于卷积和预测的快速背包算法
Pub Date : 2018-06-20 DOI: 10.1145/3188745.3188876
M. Bateni, M. Hajiaghayi, Saeed Seddighin, Cliff Stein
The knapsack problem is a fundamental problem in combinatorial optimization. It has been studied extensively from theoretical as well as practical perspectives as it is one of the most well-known NP-hard problems. The goal is to pack a knapsack of size t with the maximum value from a collection of n items with given sizes and values. Recent evidence suggests that a classic O(nt) dynamic-programming solution for the knapsack problem might be the fastest in the worst case. In fact, solving the knapsack problem was shown to be computationally equivalent to the (min, +) convolution problem, which is thought to be facing a quadratic-time barrier. This hardness is in contrast to the more famous (+, ·) convolution (generally known as polynomial multiplication), that has an O(nlogn)-time solution via Fast Fourier Transform. Our main results are algorithms with near-linear running times (in terms of the size of the knapsack and the number of items) for the knapsack problem, if either the values or sizes of items are small integers. More specifically, if item sizes are integers bounded by , the running time of our algorithm is Õ((n+t)). If the item values are integers bounded by , our algorithm runs in time Õ(n+t). Best previously known running times were O(nt), O(n2) and O(n) (Pisinger, J. of Alg., 1999). At the core of our algorithms lies the prediction technique: Roughly speaking, this new technique enables us to compute the convolution of two vectors in time (n) when an approximation of the solution within an additive error of is available. Our results also improve the best known strongly polynomial time solutions for knapsack. In the limited size setting, when the items have multiplicities, the fastest strongly polynomial time algorithms for knapsack run in time O(n2 2) and O(n3 2) for the cases of infinite and given multiplicities, respectively. Our results improve both running times by a factor of (n max{1, n/}).
背包问题是组合优化中的一个基本问题。由于它是最著名的np困难问题之一,从理论和实践的角度对它进行了广泛的研究。目标是用给定大小和值的n个物品集合中的最大值打包一个大小为t的背包。最近的证据表明,在最坏的情况下,背包问题的经典O(nt)动态规划解决方案可能是最快的。事实上,解决背包问题被证明在计算上等同于(min, +)卷积问题,后者被认为面临二次时间障碍。这种硬度与更著名的(+,·)卷积(通常称为多项式乘法)形成对比,后者通过快速傅里叶变换具有O(nlogn)时间解决方案。如果物品的值或大小是小整数,我们的主要结果是具有近似线性运行时间(就背包的大小和物品的数量而言)的算法。更具体地说,如果项目大小是整数,则算法的运行时间为Õ((n+t))。如果项值是有边界的整数,我们的算法运行时间为Õ(n+t)。以前已知的最佳运行时间是O(nt)、O(n2)和O(n) (Pisinger, J. of Alg)。, 1999)。我们算法的核心是预测技术:粗略地说,这种新技术使我们能够计算两个向量在时间(n)上的卷积,当在可加性误差范围内的近似解可用时。我们的结果也改进了最著名的强多项式时间解。在有限大小的设置中,当项目具有多重性时,对于无限多重性和给定多重性的情况,最快的强多项式时间背包算法分别在O(n2 2)和O(n32)时间内运行。我们的结果将运行时间提高了(n max{1, n/})倍。
{"title":"Fast algorithms for knapsack via convolution and prediction","authors":"M. Bateni, M. Hajiaghayi, Saeed Seddighin, Cliff Stein","doi":"10.1145/3188745.3188876","DOIUrl":"https://doi.org/10.1145/3188745.3188876","url":null,"abstract":"The knapsack problem is a fundamental problem in combinatorial optimization. It has been studied extensively from theoretical as well as practical perspectives as it is one of the most well-known NP-hard problems. The goal is to pack a knapsack of size t with the maximum value from a collection of n items with given sizes and values. Recent evidence suggests that a classic O(nt) dynamic-programming solution for the knapsack problem might be the fastest in the worst case. In fact, solving the knapsack problem was shown to be computationally equivalent to the (min, +) convolution problem, which is thought to be facing a quadratic-time barrier. This hardness is in contrast to the more famous (+, ·) convolution (generally known as polynomial multiplication), that has an O(nlogn)-time solution via Fast Fourier Transform. Our main results are algorithms with near-linear running times (in terms of the size of the knapsack and the number of items) for the knapsack problem, if either the values or sizes of items are small integers. More specifically, if item sizes are integers bounded by , the running time of our algorithm is Õ((n+t)). If the item values are integers bounded by , our algorithm runs in time Õ(n+t). Best previously known running times were O(nt), O(n2) and O(n) (Pisinger, J. of Alg., 1999). At the core of our algorithms lies the prediction technique: Roughly speaking, this new technique enables us to compute the convolution of two vectors in time (n) when an approximation of the solution within an additive error of is available. Our results also improve the best known strongly polynomial time solutions for knapsack. In the limited size setting, when the items have multiplicities, the fastest strongly polynomial time algorithms for knapsack run in time O(n2 2) and O(n3 2) for the cases of infinite and given multiplicities, respectively. Our results improve both running times by a factor of (n max{1, n/}).","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89220302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
期刊
Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1