首页 > 最新文献

Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing最新文献

英文 中文
On the complexity of hazard-free circuits 关于无危险电路的复杂性
Pub Date : 2017-11-06 DOI: 10.1145/3188745.3188912
Christian Ikenmeyer, Balagopal Komarath, C. Lenzen, Vladimir Lysikov, A. Mokhov, Karteek Sreenivasaiah
The problem of constructing hazard-free Boolean circuits dates back to the 1940s and is an important problem in circuit design. Our main lower-bound result unconditionally shows the existence of functions whose circuit complexity is polynomially bounded while every hazard-free implementation is provably of exponential size. Previous lower bounds on the hazard-free complexity were only valid for depth 2 circuits. The same proof method yields that every subcubic implementation of Boolean matrix multiplication must have hazards. These results follow from a crucial structural insight: Hazard-free complexity is a natural generalization of monotone complexity to all (not necessarily monotone) Boolean functions. Thus, we can apply known monotone complexity lower bounds to find lower bounds on the hazard-free complexity. We also lift these methods from the monotone setting to prove exponential hazard-free complexity lower bounds for non-monotone functions. As our main upper-bound result we show how to efficiently convert a Boolean circuit into a bounded-bit hazard-free circuit with only a polynomially large blow-up in the number of gates. Previously, the best known method yielded exponentially large circuits in the worst case, so our algorithm gives an exponential improvement. As a side result we establish the NP-completeness of several hazard detection problems.
构造无危险布尔电路的问题可以追溯到20世纪40年代,是电路设计中的一个重要问题。我们的主要下界结果无条件地证明了电路复杂度是多项式有界的函数的存在性,而每个无害化实现都是指数大小的。之前关于无危险复杂度的下界只对深度为2的电路有效。同样的证明方法得出,布尔矩阵乘法的每一个次立方实现都有危险。这些结果来自于一个关键的结构洞察力:无危险复杂性是单调复杂性对所有(不一定是单调的)布尔函数的自然推广。因此,我们可以应用已知的单调复杂度下界来求无危险复杂度的下界。我们还将这些方法从单调集合中提出来,证明了非单调函数的指数无危险复杂度下界。作为我们的主要上界结果,我们展示了如何有效地将布尔电路转换为有界位无危险电路,仅在门的数量上出现多项式大的爆炸。以前,最著名的方法在最坏的情况下产生指数级的大电路,因此我们的算法给出了指数级的改进。作为附带结果,我们建立了几个危害检测问题的np完备性。
{"title":"On the complexity of hazard-free circuits","authors":"Christian Ikenmeyer, Balagopal Komarath, C. Lenzen, Vladimir Lysikov, A. Mokhov, Karteek Sreenivasaiah","doi":"10.1145/3188745.3188912","DOIUrl":"https://doi.org/10.1145/3188745.3188912","url":null,"abstract":"The problem of constructing hazard-free Boolean circuits dates back to the 1940s and is an important problem in circuit design. Our main lower-bound result unconditionally shows the existence of functions whose circuit complexity is polynomially bounded while every hazard-free implementation is provably of exponential size. Previous lower bounds on the hazard-free complexity were only valid for depth 2 circuits. The same proof method yields that every subcubic implementation of Boolean matrix multiplication must have hazards. These results follow from a crucial structural insight: Hazard-free complexity is a natural generalization of monotone complexity to all (not necessarily monotone) Boolean functions. Thus, we can apply known monotone complexity lower bounds to find lower bounds on the hazard-free complexity. We also lift these methods from the monotone setting to prove exponential hazard-free complexity lower bounds for non-monotone functions. As our main upper-bound result we show how to efficiently convert a Boolean circuit into a bounded-bit hazard-free circuit with only a polynomially large blow-up in the number of gates. Previously, the best known method yielded exponentially large circuits in the worst case, so our algorithm gives an exponential improvement. As a side result we establish the NP-completeness of several hazard detection problems.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"28 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83730360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Nearly work-efficient parallel algorithm for digraph reachability 有向图可达性的近乎高效并行算法
Pub Date : 2017-11-06 DOI: 10.1145/3188745.3188926
Jeremy T. Fineman
One of the simplest problems on directed graphs is that of identifying the set of vertices reachable from a designated source vertex. This problem can be solved easily sequentially by performing a graph search, but efficient parallel algorithms have eluded researchers for decades. For sparse high-diameter graphs in particular, there is no known work-efficient parallel algorithm with nontrivial parallelism. This amounts to one of the most fundamental open questions in parallel graph algorithms: Is there a parallel algorithm for digraph reachability with nearly linear work? This paper shows that the answer is yes. This paper presents a randomized parallel algorithm for digraph reachability and related problems with expected work Õ(m) and span Õ(n2/3), and hence parallelism Ω(m/n2/3) = Ω(n1/3), on any graph with n vertices and m arcs. This is the first parallel algorithm having both nearly linear work and strongly sublinear span, i.e., span Õ(n1−є) for any constant є>0. The algorithm can be extended to produce a directed spanning tree, determine whether the graph is acyclic, topologically sort the strongly connected components of the graph, or produce a directed ear decomposition, all with work Õ(m) and span Õ(n2/3). The main technical contribution is an efficient Monte Carlo algorithm that, through the addition of Õ(n) shortcuts, reduces the diameter of the graph to Õ(n2/3) with high probability. While both sequential and parallel algorithms are known with those combinatorial properties, even the sequential algorithms are not efficient, having sequential runtime Ω(mnΩ(1)). This paper presents a surprisingly simple sequential algorithm that achieves the stated diameter reduction and runs in Õ(m) time. Parallelizing that algorithm yields the main result, but doing so involves overcoming several other challenges.
有向图中最简单的问题之一是确定从指定的源顶点可到达的顶点集。这个问题可以通过执行图搜索轻松地顺序解决,但有效的并行算法已经困扰了研究人员几十年。特别是对于稀疏的大直径图,没有已知的具有非平凡并行性的高效并行算法。这相当于并行图算法中最基本的开放问题之一:是否存在近似线性工作的有向图可达性并行算法?本文表明,答案是肯定的。本文提出了一种随机并行算法,用于求解有向图可达性和相关问题,其期望功Õ(m)和跨度Õ(n2/3),因此并行度Ω(m/n2/3) = Ω(n1/3),适用于任何有n个顶点和m条弧的图。这是第一个同时具有近线性工作和强次线性张成空间的并行算法,即对于任何常数n >0,张成空间Õ(n1−n)。该算法可以扩展为生成有向生成树,确定图是否为无环,对图的强连接分量进行拓扑排序,或生成有向耳分解,所有这些都具有功Õ(m)和跨度Õ(n2/3)。主要的技术贡献是一种高效的蒙特卡罗算法,通过添加Õ(n)捷径,高概率地将图的直径减小到Õ(n2/3)。虽然顺序算法和并行算法都具有这些组合属性,但即使是顺序算法也不高效,因为它们具有顺序运行时Ω(mnΩ(1))。本文提出了一个非常简单的顺序算法,实现了规定的直径缩小,运行时间为Õ(m)。并行化该算法会产生主要结果,但这样做需要克服其他几个挑战。
{"title":"Nearly work-efficient parallel algorithm for digraph reachability","authors":"Jeremy T. Fineman","doi":"10.1145/3188745.3188926","DOIUrl":"https://doi.org/10.1145/3188745.3188926","url":null,"abstract":"One of the simplest problems on directed graphs is that of identifying the set of vertices reachable from a designated source vertex. This problem can be solved easily sequentially by performing a graph search, but efficient parallel algorithms have eluded researchers for decades. For sparse high-diameter graphs in particular, there is no known work-efficient parallel algorithm with nontrivial parallelism. This amounts to one of the most fundamental open questions in parallel graph algorithms: Is there a parallel algorithm for digraph reachability with nearly linear work? This paper shows that the answer is yes. This paper presents a randomized parallel algorithm for digraph reachability and related problems with expected work Õ(m) and span Õ(n2/3), and hence parallelism Ω(m/n2/3) = Ω(n1/3), on any graph with n vertices and m arcs. This is the first parallel algorithm having both nearly linear work and strongly sublinear span, i.e., span Õ(n1−є) for any constant є>0. The algorithm can be extended to produce a directed spanning tree, determine whether the graph is acyclic, topologically sort the strongly connected components of the graph, or produce a directed ear decomposition, all with work Õ(m) and span Õ(n2/3). The main technical contribution is an efficient Monte Carlo algorithm that, through the addition of Õ(n) shortcuts, reduces the diameter of the graph to Õ(n2/3) with high probability. While both sequential and parallel algorithms are known with those combinatorial properties, even the sequential algorithms are not efficient, having sequential runtime Ω(mnΩ(1)). This paper presents a surprisingly simple sequential algorithm that achieves the stated diameter reduction and runs in Õ(m) time. Parallelizing that algorithm yields the main result, but doing so involves overcoming several other challenges.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82798122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Capacity upper bounds for deletion-type channels 删除类型通道的容量上限
Pub Date : 2017-11-05 DOI: 10.1145/3188745.3188768
Mahdi Cheraghchi
We develop a systematic approach, based on convex programming and real analysis, for obtaining upper bounds on the capacity of the binary deletion channel and, more generally, channels with i.i.d. insertions and deletions. Other than the classical deletion channel, we give a special attention to the Poisson-repeat channel introduced by Mitzenmacher and Drinea (IEEE Transactions on Information Theory, 2006). Our framework can be applied to obtain capacity upper bounds for any repetition distribution (the deletion and Poisson-repeat channels corresponding to the special cases of Bernoulli and Poisson distributions). Our techniques essentially reduce the task of proving capacity upper bounds to maximizing a univariate, real-valued, and often concave function over a bounded interval. The corresponding univariate function is carefully designed according to the underlying distribution of repetitions and the choices vary depending on the desired strength of the upper bounds as well as the desired simplicity of the function (e.g., being only efficiently computable versus having an explicit closed-form expression in terms of elementary, or common special, functions). Among our results, we show that the capacity of the binary deletion channel with deletion probability d is at most (1−d) logϕ for d ≥ 1/2, and, assuming the capacity function is convex, is at most 1−d log(4/ϕ) for d<1/2, where ϕ=(1+√5)/2 is the golden ratio. This is the first nontrivial capacity upper bound for any value of d outside the limiting case d → 0 that is fully explicit and proved without computer assistance. Furthermore, we derive the first set of capacity upper bounds for the Poisson-repeat channel. Our results uncover further striking connections between this channel and the deletion channel, and suggest, somewhat counter-intuitively, that the Poisson-repeat channel is actually analytically simpler than the deletion channel and may be of key importance to a complete understanding of the deletion channel. Finally, we derive several novel upper bounds on the capacity of the deletion channel. All upper bounds are maximums of efficiently computable, and concave, univariate real functions over a bounded domain. In turn, we upper bound these functions in terms of explicit elementary and standard special functions, whose maximums can be found even more efficiently (and sometimes, analytically, for example for d=1/2). Along the way, we develop several new techniques of potentially independent interest. For example, we develop systematic techniques to study channels with mean constraints over the reals. Furthermore, we motivate the study of novel probability distributions over non-negative integers, as well as novel special functions which could be of interest to mathematical analysis.
我们开发了一种基于凸规划和实数分析的系统方法,用于获得二进制删除信道的容量上界,更一般地说,具有iid插入和删除的信道。除了经典的删除信道,我们特别关注由Mitzenmacher和Drinea (IEEE Transactions on Information Theory, 2006)引入的泊松重复信道。我们的框架可用于获得任何重复分布(与伯努利分布和泊松分布的特殊情况相对应的删除和泊松-重复通道)的容量上界。我们的技术从本质上减少了证明容量上界的任务,使单变量、实值和通常在有界区间内凹函数最大化。相应的单变量函数是根据重复的潜在分布精心设计的,选择取决于上界的期望强度以及函数的期望简单性(例如,仅可有效计算,而不是根据初等函数或普通特殊函数具有显式的封闭形式表达式)。在我们的研究结果中,我们表明,对于d≥1/2,删除概率为d的二进制删除通道的容量最多为(1−d) logϕ,并且,假设容量函数是凸的,对于d<1/2,最大为1−d log(4/ϕ),其中ϕ=(1+√5)/2是黄金比例。这是在极限情况d→0之外的任何d值的非平凡容量上界的第一个完全显式且无需计算机辅助证明的。进一步,我们导出了泊松重复信道的第一组容量上界。我们的研究结果进一步揭示了该通道和删除通道之间的惊人联系,并表明,在某种程度上与直觉相反,泊松重复通道实际上在分析上比删除通道更简单,并且可能对完全理解删除通道至关重要。最后,我们给出了删除信道容量的几个新的上界。所有上界都是有效可计算的、凹的、单变量实函数在有界域上的最大值。反过来,我们用显式初等函数和标准特殊函数为这些函数上界,它们的最大值可以更有效地找到(有时,解析地,例如d=1/2)。在此过程中,我们开发了几种潜在的独立兴趣的新技术。例如,我们开发了系统的技术来研究具有对实数的平均约束的信道。此外,我们鼓励研究非负整数上的新概率分布,以及可能对数学分析感兴趣的新特殊函数。
{"title":"Capacity upper bounds for deletion-type channels","authors":"Mahdi Cheraghchi","doi":"10.1145/3188745.3188768","DOIUrl":"https://doi.org/10.1145/3188745.3188768","url":null,"abstract":"We develop a systematic approach, based on convex programming and real analysis, for obtaining upper bounds on the capacity of the binary deletion channel and, more generally, channels with i.i.d. insertions and deletions. Other than the classical deletion channel, we give a special attention to the Poisson-repeat channel introduced by Mitzenmacher and Drinea (IEEE Transactions on Information Theory, 2006). Our framework can be applied to obtain capacity upper bounds for any repetition distribution (the deletion and Poisson-repeat channels corresponding to the special cases of Bernoulli and Poisson distributions). Our techniques essentially reduce the task of proving capacity upper bounds to maximizing a univariate, real-valued, and often concave function over a bounded interval. The corresponding univariate function is carefully designed according to the underlying distribution of repetitions and the choices vary depending on the desired strength of the upper bounds as well as the desired simplicity of the function (e.g., being only efficiently computable versus having an explicit closed-form expression in terms of elementary, or common special, functions). Among our results, we show that the capacity of the binary deletion channel with deletion probability d is at most (1−d) logϕ for d ≥ 1/2, and, assuming the capacity function is convex, is at most 1−d log(4/ϕ) for d<1/2, where ϕ=(1+√5)/2 is the golden ratio. This is the first nontrivial capacity upper bound for any value of d outside the limiting case d → 0 that is fully explicit and proved without computer assistance. Furthermore, we derive the first set of capacity upper bounds for the Poisson-repeat channel. Our results uncover further striking connections between this channel and the deletion channel, and suggest, somewhat counter-intuitively, that the Poisson-repeat channel is actually analytically simpler than the deletion channel and may be of key importance to a complete understanding of the deletion channel. Finally, we derive several novel upper bounds on the capacity of the deletion channel. All upper bounds are maximums of efficiently computable, and concave, univariate real functions over a bounded domain. In turn, we upper bound these functions in terms of explicit elementary and standard special functions, whose maximums can be found even more efficiently (and sometimes, analytically, for example for d=1/2). Along the way, we develop several new techniques of potentially independent interest. For example, we develop systematic techniques to study channels with mean constraints over the reals. Furthermore, we motivate the study of novel probability distributions over non-negative integers, as well as novel special functions which could be of interest to mathematical analysis.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"69 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84894013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Constant approximation for k-median and k-means with outliers via iterative rounding 通过迭代舍入对具有异常值的k中值和k均值进行常数逼近
Pub Date : 2017-11-03 DOI: 10.1145/3188745.3188882
Ravishankar Krishnaswamy, Shi Li, Sai Sandeep
In this paper, we present a new iterative rounding framework for many clustering problems. Using this, we obtain an (α1 + є ≤ 7.081 + є)-approximation algorithm for k-median with outliers, greatly improving upon the large implicit constant approximation ratio of Chen. For k-means with outliers, we give an (α2+є ≤ 53.002 + є)-approximation, which is the first O(1)-approximation for this problem. The iterative algorithm framework is very versatile; we show how it can be used to give α1- and (α1 + є)-approximation algorithms for matroid and knapsack median problems respectively, improving upon the previous best approximations ratios of 8 due to Swamy and 17.46 due to Byrka et al. The natural LP relaxation for the k-median/k-means with outliers problem has an unbounded integrality gap. In spite of this negative result, our iterative rounding framework shows that we can round an LP solution to an almost-integral solution of small cost, in which we have at most two fractionally open facilities. Thus, the LP integrality gap arises due to the gap between almost-integral and fully-integral solutions. Then, using a pre-processing procedure, we show how to convert an almost-integral solution to a fully-integral solution losing only a constant-factor in the approximation ratio. By further using a sparsification technique, the additive factor loss incurred by the conversion can be reduced to any є > 0.
在本文中,我们提出了一个新的迭代舍入框架,用于许多聚类问题。在此基础上,我们得到了具有异常值的k-中位数的(α1 + k≤7.081 + k)近似算法,大大改进了Chen的大隐式常数近似比。对于带有异常值的k-means,我们给出了一个(α2+ tu≤53.002 + tu)-近似,这是该问题的第一个O(1)-近似。迭代算法框架是非常通用的;我们展示了如何使用它分别给出矩阵和背包中值问题的α1-和(α1 + -)-近似算法,改进了先前由Swamy和Byrka等人给出的最佳近似比率为8和17.46。带离群值的k-中值/k-均值问题的自然LP松弛具有无界的完整性缺口。尽管有这个负面的结果,我们的迭代舍入框架表明,我们可以将LP解舍入为一个小成本的几乎积分解,其中我们最多有两个部分开放的设施。因此,由于几乎整解和完全整解之间的差距,产生了LP完整性差距。然后,使用预处理程序,我们展示了如何将几乎积分解转换为完全积分解,仅损失近似比率中的常数因子。通过进一步使用稀疏化技术,转换引起的加性因子损失可以减小到任意> 0。
{"title":"Constant approximation for k-median and k-means with outliers via iterative rounding","authors":"Ravishankar Krishnaswamy, Shi Li, Sai Sandeep","doi":"10.1145/3188745.3188882","DOIUrl":"https://doi.org/10.1145/3188745.3188882","url":null,"abstract":"In this paper, we present a new iterative rounding framework for many clustering problems. Using this, we obtain an (α1 + є ≤ 7.081 + є)-approximation algorithm for k-median with outliers, greatly improving upon the large implicit constant approximation ratio of Chen. For k-means with outliers, we give an (α2+є ≤ 53.002 + є)-approximation, which is the first O(1)-approximation for this problem. The iterative algorithm framework is very versatile; we show how it can be used to give α1- and (α1 + є)-approximation algorithms for matroid and knapsack median problems respectively, improving upon the previous best approximations ratios of 8 due to Swamy and 17.46 due to Byrka et al. The natural LP relaxation for the k-median/k-means with outliers problem has an unbounded integrality gap. In spite of this negative result, our iterative rounding framework shows that we can round an LP solution to an almost-integral solution of small cost, in which we have at most two fractionally open facilities. Thus, the LP integrality gap arises due to the gap between almost-integral and fully-integral solutions. Then, using a pre-processing procedure, we show how to convert an almost-integral solution to a fully-integral solution losing only a constant-factor in the approximation ratio. By further using a sparsification technique, the additive factor loss incurred by the conversion can be reduced to any є > 0.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"240 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75578578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 97
An homotopy method for lp regression provably beyond self-concordance and in input-sparsity time 在输入稀疏时间内证明lp回归的一种超越自协调的同伦方法
Pub Date : 2017-11-03 DOI: 10.1145/3188745.3188776
Sébastien Bubeck, Michael B. Cohen, Y. Lee, Yuanzhi Li
We consider the problem of linear regression where the ℓ2n norm loss (i.e., the usual least squares loss) is replaced by the ℓpn norm. We show how to solve such problems up to machine precision in Õp(n|1/2 − 1/p|) (dense) matrix-vector products and Õp(1) matrix inversions, or alternatively in Õp(n|1/2 − 1/p|) calls to a (sparse) linear system solver. This improves the state of the art for any p∉{1,2,+∞}. Furthermore we also propose a randomized algorithm solving such problems in input sparsity time, i.e., Õp(N + poly(d)) where N is the size of the input and d is the number of variables. Such a result was only known for p=2. Finally we prove that these results lie outside the scope of the Nesterov-Nemirovski’s theory of interior point methods by showing that any symmetric self-concordant barrier on the ℓpn unit ball has self-concordance parameter Ω(n).
我们考虑线性回归的问题,其中l2n范数损失(即,通常的最小二乘损失)被lpn范数取代。我们展示了如何在Õp(n|1/2−1/p|)(密集)矩阵向量积和Õp(1)矩阵反转中解决这些问题,或者在Õp(n|1/2−1/p|)调用(稀疏)线性系统求解器中解决这些问题。这提高了任意p∈{1,2,+∞}的技术水平。此外,我们还提出了一个在输入稀疏时间内解决此类问题的随机算法,即Õp(N + poly(d)),其中N是输入的大小,d是变量的数量。只有在p=2时才知道这样的结果。最后,我们证明了这些结果不在Nesterov-Nemirovski的内点法理论的范围之内,证明了在1 pn单位球上的任何对称自协势垒都有自协参数Ω(n)。
{"title":"An homotopy method for lp regression provably beyond self-concordance and in input-sparsity time","authors":"Sébastien Bubeck, Michael B. Cohen, Y. Lee, Yuanzhi Li","doi":"10.1145/3188745.3188776","DOIUrl":"https://doi.org/10.1145/3188745.3188776","url":null,"abstract":"We consider the problem of linear regression where the ℓ2n norm loss (i.e., the usual least squares loss) is replaced by the ℓpn norm. We show how to solve such problems up to machine precision in Õp(n|1/2 − 1/p|) (dense) matrix-vector products and Õp(1) matrix inversions, or alternatively in Õp(n|1/2 − 1/p|) calls to a (sparse) linear system solver. This improves the state of the art for any p∉{1,2,+∞}. Furthermore we also propose a randomized algorithm solving such problems in input sparsity time, i.e., Õp(N + poly(d)) where N is the size of the input and d is the number of variables. Such a result was only known for p=2. Finally we prove that these results lie outside the scope of the Nesterov-Nemirovski’s theory of interior point methods by showing that any symmetric self-concordant barrier on the ℓpn unit ball has self-concordance parameter Ω(n).","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"19 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86017235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 49
k-server via multiscale entropic regularization 基于多尺度熵正则化的K-server
Pub Date : 2017-11-03 DOI: 10.1145/3188745.3188798
Sébastien Bubeck, Michael B. Cohen, James R. Lee, Y. Lee, A. Madry
We present an O((logk)2)-competitive randomized algorithm for the k-server problem on hierarchically separated trees (HSTs). This is the first o(k)-competitive randomized algorithm for which the competitive ratio is independent of the size of the underlying HST. Our algorithm is designed in the framework of online mirror descent where the mirror map is a multiscale entropy. When combined with Bartal’s static HST embedding reduction, this leads to an O((logk)2 logn)-competitive algorithm on any n-point metric space. We give a new dynamic HST embedding that yields an O((logk)3 logΔ)-competitive algorithm on any metric space where the ratio of the largest to smallest non-zero distance is at most Δ.
提出了一种O((logk)2)竞争随机化算法,用于求解层次分离树(HSTs)上的k-server问题。这是第一个0 (k)竞争随机算法,其竞争比率与底层HST的大小无关。我们的算法是在镜像下降的框架下设计的,其中镜像映射是一个多尺度熵。当与Bartal的静态HST嵌入约简相结合时,这导致了在任何n点度量空间上的O((logk)2 logn)竞争算法。我们给出了一种新的动态HST嵌入,它在任何度量空间上产生O((logk)3 logΔ)竞争算法,其中最大与最小非零距离之比最多为Δ。
{"title":"k-server via multiscale entropic regularization","authors":"Sébastien Bubeck, Michael B. Cohen, James R. Lee, Y. Lee, A. Madry","doi":"10.1145/3188745.3188798","DOIUrl":"https://doi.org/10.1145/3188745.3188798","url":null,"abstract":"We present an O((logk)2)-competitive randomized algorithm for the k-server problem on hierarchically separated trees (HSTs). This is the first o(k)-competitive randomized algorithm for which the competitive ratio is independent of the size of the underlying HST. Our algorithm is designed in the framework of online mirror descent where the mirror map is a multiscale entropy. When combined with Bartal’s static HST embedding reduction, this leads to an O((logk)2 logn)-competitive algorithm on any n-point metric space. We give a new dynamic HST embedding that yields an O((logk)3 logΔ)-competitive algorithm on any metric space where the ratio of the largest to smallest non-zero distance is at most Δ.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"80 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90685912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 81
Shadow tomography of quantum states 量子态的影层析成像
Pub Date : 2017-11-03 DOI: 10.1145/3188745.3188802
S. Aaronson
We introduce the problem of *shadow tomography*: given an unknown D-dimensional quantum mixed state ρ, as well as known two-outcome measurements E1,…,EM, estimate the probability that Ei accepts ρ, to within additive error ε, for each of the M measurements. How many copies of ρ are needed to achieve this, with high probability? Surprisingly, we give a procedure that solves the problem by measuring only O( ε−5·log4 M·logD) copies. This means, for example, that we can learn the behavior of an arbitrary n-qubit state, on *all* accepting/rejecting circuits of some fixed polynomial size, by measuring only nO( 1) copies of the state. This resolves an open problem of the author, which arose from his work on private-key quantum money schemes, but which also has applications to quantum copy-protected software, quantum advice, and quantum one-way communication. Recently, building on this work, Brandão et al. have given a different approach to shadow tomography using semidefinite programming, which achieves a savings in computation time.
我们引入了阴影层析成像的问题:给定一个未知的d维量子混合态ρ,以及已知的双结果测量E1,…,EM,估计Ei接受ρ的概率,在加性误差ε范围内,对于每一个M测量。高概率地,需要多少个ρ的拷贝?令人惊讶的是,我们给出了一个仅通过测量O(ε−5·log4 M·logD)拷贝来解决问题的程序。这意味着,例如,我们可以通过测量状态的nO(1)个拷贝来学习任意n-量子比特状态的行为,在某个固定多项式大小的*所有*接受/拒绝电路上。这解决了作者的一个开放性问题,这个问题来自于他对私钥量子货币方案的研究,但它也适用于量子复制保护软件、量子建议和量子单向通信。最近,在这项工作的基础上,brand等人使用半定规划给出了一种不同的阴影层析成像方法,从而节省了计算时间。
{"title":"Shadow tomography of quantum states","authors":"S. Aaronson","doi":"10.1145/3188745.3188802","DOIUrl":"https://doi.org/10.1145/3188745.3188802","url":null,"abstract":"We introduce the problem of *shadow tomography*: given an unknown D-dimensional quantum mixed state ρ, as well as known two-outcome measurements E1,…,EM, estimate the probability that Ei accepts ρ, to within additive error ε, for each of the M measurements. How many copies of ρ are needed to achieve this, with high probability? Surprisingly, we give a procedure that solves the problem by measuring only O( ε−5·log4 M·logD) copies. This means, for example, that we can learn the behavior of an arbitrary n-qubit state, on *all* accepting/rejecting circuits of some fixed polynomial size, by measuring only nO( 1) copies of the state. This resolves an open problem of the author, which arose from his work on private-key quantum money schemes, but which also has applications to quantum copy-protected software, quantum advice, and quantum one-way communication. Recently, building on this work, Brandão et al. have given a different approach to shadow tomography using semidefinite programming, which achieves a savings in computation time.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"491 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77808562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 280
A simply exponential upper bound on the maximum number of stable matchings 稳定匹配的最大数目的简单指数上界
Pub Date : 2017-11-03 DOI: 10.1145/3188745.3188848
Anna R. Karlin, S. Gharan, Robbie Weber
Stable matching is a classical combinatorial problem that has been the subject of intense theoretical and empirical study since its introduction in 1962 in a seminal paper by Gale and Shapley. In this paper, we provide a new upper bound on f(n), the maximum number of stable matchings that a stable matching instance with n men and n women can have. It has been a long-standing open problem to understand the asymptotic behavior of f(n) as n→∞, first posed by Donald Knuth in the 1970s. Until now the best lower bound was approximately 2.28n, and the best upper bound was 2nlogn− O(n). In this paper, we show that for all n, f(n) ≤ cn for some universal constant c. This matches the lower bound up to the base of the exponent. Our proof is based on a reduction to counting the number of downsets of a family of posets that we call “mixing”. The latter might be of independent interest.
稳定匹配是一个经典的组合问题,自1962年Gale和Shapley在一篇开创性的论文中提出以来,一直是激烈的理论和实证研究的主题。本文给出了一个新的上界f(n),即一个有n个男人和n个女人的稳定匹配实例的最大稳定匹配数。理解f(n)在n→∞时的渐近行为是一个长期存在的开放性问题,最早是由Donald Knuth在20世纪70年代提出的。到目前为止,最佳下界约为2.28n,最佳上界为2nlogn−O(n)。在本文中,我们证明了对于所有n, f(n)≤cn对于某个普适常数c,它匹配到指数底的下界。我们的证明是基于对一系列我们称之为“混合”的偏序集的下集数量的计算。后者可能具有独立的利益。
{"title":"A simply exponential upper bound on the maximum number of stable matchings","authors":"Anna R. Karlin, S. Gharan, Robbie Weber","doi":"10.1145/3188745.3188848","DOIUrl":"https://doi.org/10.1145/3188745.3188848","url":null,"abstract":"Stable matching is a classical combinatorial problem that has been the subject of intense theoretical and empirical study since its introduction in 1962 in a seminal paper by Gale and Shapley. In this paper, we provide a new upper bound on f(n), the maximum number of stable matchings that a stable matching instance with n men and n women can have. It has been a long-standing open problem to understand the asymptotic behavior of f(n) as n→∞, first posed by Donald Knuth in the 1970s. Until now the best lower bound was approximately 2.28n, and the best upper bound was 2nlogn− O(n). In this paper, we show that for all n, f(n) ≤ cn for some universal constant c. This matches the lower bound up to the base of the exponent. Our proof is based on a reduction to counting the number of downsets of a family of posets that we call “mixing”. The latter might be of independent interest.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"111 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85815321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Inapproximability of the independent set polynomial in the complex plane 复平面上独立集多项式的不可逼近性
Pub Date : 2017-11-01 DOI: 10.1145/3188745.3188788
Ivona Bezáková, Andreas Galanis, L. A. Goldberg, Daniel Stefankovic
We study the complexity of approximating the value of the independent set polynomial ZG(λ) of a graph G with maximum degree Δ when the activity λ is a complex number. When λ is real, the complexity picture is well-understood, and is captured by two real-valued thresholds λ* and λc, which depend on Δ and satisfy 0<λ*<λc. It is known that if λ is a real number in the interval (−λ*,λc) then there is an FPTAS for approximating ZG(λ) on graphs G with maximum degree at most Δ. On the other hand, if λ is a real number outside of the (closed) interval, then approximation is NP-hard. The key to establishing this picture was the interpretation of the thresholds λ* and λc on the Δ-regular tree. The ”occupation ratio” of a Δ-regular tree T is the contribution to ZT(λ) from independent sets containing the root of the tree, divided by ZT(λ) itself. This occupation ratio converges to a limit, as the height of the tree grows, if and only if λ∈ [−λ*,λc]. Unsurprisingly, the case where λ is complex is more challenging. It is known that there is an FPTAS when λ is a complex number with norm at most λ* and also when λ is in a small strip surrounding the real interval [0,λc). However, neither of these results is believed to fully capture the truth about when approximation is possible. Peters and Regts identified the values of λ for which the occupation ratio of the Δ-regular tree converges. These values carve a cardioid-shaped region ΛΔ in the complex plane, whose boundary includes the critical points −λ* and λc. Motivated by the picture in the real case, they asked whether ΛΔ marks the true approximability threshold for general complex values λ. Our main result shows that for every λ outside of ΛΔ, the problem of approximating ZG(λ) on graphs G with maximum degree at most Δ is indeed NP-hard. In fact, when λ is outside of ΛΔ and is not a positive real number, we give the stronger result that approximating ZG(λ) is actually #P-hard. Further, on the negative real axis, when λ<−λ*, we show that it is #P-hard to even decide whether ZG(λ)>0, resolving in the affirmative a conjecture of Harvey, Srivastava and Vondrak. Our proof techniques are based around tools from complex analysis — specifically the study of iterative multivariate rational maps.
研究了当活度λ为复数时,最大度为Δ的图G的独立集多项式ZG(λ)的逼近复杂度。当λ为实数时,复杂性图像被很好地理解,并由两个实值阈值λ*和λc捕获,它们依赖于Δ并满足00,肯定地解决了Harvey, Srivastava和Vondrak的猜想。我们的证明技术是基于复杂分析的工具-特别是迭代多元理性映射的研究。
{"title":"Inapproximability of the independent set polynomial in the complex plane","authors":"Ivona Bezáková, Andreas Galanis, L. A. Goldberg, Daniel Stefankovic","doi":"10.1145/3188745.3188788","DOIUrl":"https://doi.org/10.1145/3188745.3188788","url":null,"abstract":"We study the complexity of approximating the value of the independent set polynomial ZG(λ) of a graph G with maximum degree Δ when the activity λ is a complex number. When λ is real, the complexity picture is well-understood, and is captured by two real-valued thresholds λ* and λc, which depend on Δ and satisfy 0<λ*<λc. It is known that if λ is a real number in the interval (−λ*,λc) then there is an FPTAS for approximating ZG(λ) on graphs G with maximum degree at most Δ. On the other hand, if λ is a real number outside of the (closed) interval, then approximation is NP-hard. The key to establishing this picture was the interpretation of the thresholds λ* and λc on the Δ-regular tree. The ”occupation ratio” of a Δ-regular tree T is the contribution to ZT(λ) from independent sets containing the root of the tree, divided by ZT(λ) itself. This occupation ratio converges to a limit, as the height of the tree grows, if and only if λ∈ [−λ*,λc]. Unsurprisingly, the case where λ is complex is more challenging. It is known that there is an FPTAS when λ is a complex number with norm at most λ* and also when λ is in a small strip surrounding the real interval [0,λc). However, neither of these results is believed to fully capture the truth about when approximation is possible. Peters and Regts identified the values of λ for which the occupation ratio of the Δ-regular tree converges. These values carve a cardioid-shaped region ΛΔ in the complex plane, whose boundary includes the critical points −λ* and λc. Motivated by the picture in the real case, they asked whether ΛΔ marks the true approximability threshold for general complex values λ. Our main result shows that for every λ outside of ΛΔ, the problem of approximating ZG(λ) on graphs G with maximum degree at most Δ is indeed NP-hard. In fact, when λ is outside of ΛΔ and is not a positive real number, we give the stronger result that approximating ZG(λ) is actually #P-hard. Further, on the negative real axis, when λ<−λ*, we show that it is #P-hard to even decide whether ZG(λ)>0, resolving in the affirmative a conjecture of Harvey, Srivastava and Vondrak. Our proof techniques are based around tools from complex analysis — specifically the study of iterative multivariate rational maps.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"25 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84796726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
At the roots of dictionary compression: string attractors 字典压缩的根源:字符串吸引子
Pub Date : 2017-10-30 DOI: 10.1145/3188745.3188814
Dominik Kempa, N. Prezza
A well-known fact in the field of lossless text compression is that high-order entropy is a weak model when the input contains long repetitions. Motivated by this fact, decades of research have generated myriads of so-called dictionary compressors: algorithms able to reduce the text’s size by exploiting its repetitiveness. Lempel-Ziv 77 is one of the most successful and well-known tools of this kind, followed by straight-line programs, run-length Burrows-Wheeler transform, macro schemes, collage systems, and the compact directed acyclic word graph. In this paper, we show that these techniques are different solutions to the same, elegant, combinatorial problem: to find a small set of positions capturing all distinct text’s substrings. We call such a set a string attractor. We first show reductions between dictionary compressors and string attractors. This gives the approximation ratios of dictionary compressors with respect to the smallest string attractor and allows us to uncover new asymptotic relations between the output sizes of different dictionary compressors. We then show that the k-attractor problem — deciding whether a text has a size-t set of positions capturing all substrings of length at most k — is NP-complete for k≥ 3. This, in particular, includes the full string attractor problem. We provide several approximation techniques for the smallest k-attractor, show that the problem is APX-complete for constant k, and give strong inapproximability results. To conclude, we provide matching lower and upper bounds for the random access problem on string attractors. The upper bound is proved by showing a data structure supporting queries in optimal time. Our data structure is universal: by our reductions to string attractors, it supports random access on any dictionary-compression scheme. In particular, it matches the lower bound also on LZ77, straight-line programs, collage systems, and macro schemes, and therefore essentially closes (at once) the random access problem for all these compressors.
在无损文本压缩领域中一个众所周知的事实是,当输入包含长重复时,高阶熵是一个弱模型。在这一事实的推动下,几十年的研究已经产生了无数所谓的字典压缩器:能够通过利用文本的重复来减少文本大小的算法。Lempel-Ziv 77是这类工具中最成功和最著名的工具之一,紧随其后的是直线规划、运行长度Burrows-Wheeler变换、宏观方案、拼贴系统和紧致有向无环字图。在本文中,我们展示了这些技术是同一个优雅的组合问题的不同解决方案:找到捕获所有不同文本子字符串的一小组位置。我们称这样的集合为弦吸引子。我们首先展示了字典压缩器和字符串吸引器之间的约简。这给出了字典压缩器相对于最小字符串吸引子的近似比率,并允许我们揭示不同字典压缩器输出大小之间的新渐近关系。然后我们证明了k吸引子问题——决定一个文本是否有一个大小为t的位置集合,捕获了长度不超过k的所有子串——对于k≥3是np完全的。特别地,这包括了全弦吸引子问题。我们提供了几种最小k吸引子的逼近方法,证明了对于常数k问题是apx完备的,并给出了强的不可逼近性结果。最后,我们给出了字符串吸引子随机存取问题的匹配下界和上界。通过展示在最优时间内支持查询的数据结构来证明上界。我们的数据结构是通用的:通过对字符串吸引子的简化,它支持对任何字典压缩方案的随机访问。特别是,它还匹配LZ77、直线程序、拼贴系统和宏方案的下界,因此基本上(立即)解决了所有这些压缩器的随机访问问题。
{"title":"At the roots of dictionary compression: string attractors","authors":"Dominik Kempa, N. Prezza","doi":"10.1145/3188745.3188814","DOIUrl":"https://doi.org/10.1145/3188745.3188814","url":null,"abstract":"A well-known fact in the field of lossless text compression is that high-order entropy is a weak model when the input contains long repetitions. Motivated by this fact, decades of research have generated myriads of so-called dictionary compressors: algorithms able to reduce the text’s size by exploiting its repetitiveness. Lempel-Ziv 77 is one of the most successful and well-known tools of this kind, followed by straight-line programs, run-length Burrows-Wheeler transform, macro schemes, collage systems, and the compact directed acyclic word graph. In this paper, we show that these techniques are different solutions to the same, elegant, combinatorial problem: to find a small set of positions capturing all distinct text’s substrings. We call such a set a string attractor. We first show reductions between dictionary compressors and string attractors. This gives the approximation ratios of dictionary compressors with respect to the smallest string attractor and allows us to uncover new asymptotic relations between the output sizes of different dictionary compressors. We then show that the k-attractor problem — deciding whether a text has a size-t set of positions capturing all substrings of length at most k — is NP-complete for k≥ 3. This, in particular, includes the full string attractor problem. We provide several approximation techniques for the smallest k-attractor, show that the problem is APX-complete for constant k, and give strong inapproximability results. To conclude, we provide matching lower and upper bounds for the random access problem on string attractors. The upper bound is proved by showing a data structure supporting queries in optimal time. Our data structure is universal: by our reductions to string attractors, it supports random access on any dictionary-compression scheme. In particular, it matches the lower bound also on LZ77, straight-line programs, collage systems, and macro schemes, and therefore essentially closes (at once) the random access problem for all these compressors.","PeriodicalId":20593,"journal":{"name":"Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing","volume":"23 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83927484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 95
期刊
Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1