Pub Date : 2023-05-23DOI: https://dl.acm.org/doi/10.1145/3587250
Jason Gaitonde, Éva Tardos
Bounding the price of anarchy, which quantifies the damage to social welfare due to selfish behavior of the participants, has been an important area of research in algorithmic game theory. Classical work on such bounds in repeated games makes the strong assumption that the subsequent rounds of the repeated games are independent beyond any influence on play from past history. This work studies such bounds in environments that themselves change due to the actions of the agents. Concretely, we consider this problem in discrete-time queuing systems, where competitive queues try to get their packets served. In this model, a queue gets to send a packet at each step to one of the servers, which will attempt to serve the oldest arriving packet, and unprocessed packets are returned to each queue. We model this as a repeated game where queues compete for the capacity of the servers, but where the state of the game evolves as the length of each queue varies.
We analyze this queuing system from multiple perspectives. As a baseline measure, we first establish precise conditions on the queuing arrival rates and service capacities that ensure all packets clear efficiently under centralized coordination. We then show that if queues strategically choose servers according to independent and stationary distributions, the system remains stable provided it would be stable under coordination with arrival rates scaled up by a factor of just (frac{e}{e-1}). Finally, we extend these results to no-regret learning dynamics: if queues use learning algorithms satisfying the no-regret property to choose servers, then the requisite factor increases to 2, and both of these bounds are tight. Both of these results require new probabilistic techniques compared to the classical price of anarchy literature and show that in such settings, no-regret learning can exhibit efficiency loss due to myopia.
{"title":"The Price of Anarchy of Strategic Queuing Systems","authors":"Jason Gaitonde, Éva Tardos","doi":"https://dl.acm.org/doi/10.1145/3587250","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3587250","url":null,"abstract":"<p>Bounding the price of anarchy, which quantifies the damage to social welfare due to selfish behavior of the participants, has been an important area of research in algorithmic game theory. Classical work on such bounds in repeated games makes the strong assumption that the subsequent rounds of the repeated games are independent beyond any influence on play from past history. This work studies such bounds in environments that themselves change due to the actions of the agents. Concretely, we consider this problem in discrete-time queuing systems, where competitive queues try to get their packets served. In this model, a queue gets to send a packet at each step to one of the servers, which will attempt to serve the oldest arriving packet, and unprocessed packets are returned to each queue. We model this as a repeated game where queues compete for the capacity of the servers, but where the state of the game evolves as the length of each queue varies.</p><p>We analyze this queuing system from multiple perspectives. As a baseline measure, we first establish precise conditions on the queuing arrival rates and service capacities that ensure all packets clear efficiently under centralized coordination. We then show that if queues strategically choose servers according to independent and stationary distributions, the system remains stable provided it would be stable under coordination with arrival rates scaled up by a factor of just (frac{e}{e-1}). Finally, we extend these results to no-regret learning dynamics: if queues use learning algorithms satisfying the no-regret property to choose servers, then the requisite factor increases to 2, and both of these bounds are tight. Both of these results require new probabilistic techniques compared to the classical price of anarchy literature and show that in such settings, no-regret learning can exhibit efficiency loss due to myopia.</p>","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"78 12","pages":""},"PeriodicalIF":2.5,"publicationDate":"2023-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138525706","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-23DOI: https://dl.acm.org/doi/10.1145/3583680
Allen Liu, Ankur Moitra
This work represents a natural coalescence of two important lines of work — learning mixtures of Gaussians and algorithmic robust statistics. In particular, we give the first provably robust algorithm for learning mixtures of any constant number of Gaussians. We require only mild assumptions on the mixing weights and that the total variation distance between components is bounded away from zero. At the heart of our algorithm is a new method for proving a type of dimension-independent polynomial identifiability — which we call robust identifiability — through applying a carefully chosen sequence of differential operations to certain generating functions that not only encode the parameters we would like to learn but also the system of polynomial equations we would like to solve. We show how the symbolic identities we derive can be directly used to analyze a natural sum-of-squares relaxation.
{"title":"Robustly Learning General Mixtures of Gaussians","authors":"Allen Liu, Ankur Moitra","doi":"https://dl.acm.org/doi/10.1145/3583680","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3583680","url":null,"abstract":"<p>This work represents a natural coalescence of two important lines of work — learning mixtures of Gaussians and algorithmic robust statistics. In particular, we give the first provably robust algorithm for learning mixtures of any constant number of Gaussians. We require only mild assumptions on the mixing weights and that the total variation distance between components is bounded away from zero. At the heart of our algorithm is a new method for proving a type of dimension-independent polynomial identifiability — which we call robust identifiability — through applying a carefully chosen sequence of differential operations to certain generating functions that not only encode the parameters we would like to learn but also the system of polynomial equations we would like to solve. We show how the symbolic identities we derive can be directly used to analyze a natural sum-of-squares relaxation.</p>","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"54 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2023-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138525678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-23DOI: https://dl.acm.org/doi/10.1145/3588866
Laurent Doyen
We consider two-player stochastic games played on a finite graph for infinitely many rounds. Stochastic games generalize both Markov decision processes (MDP) by adding an adversary player, and two-player deterministic games by adding stochasticity. The outcome of the game is a sequence of distributions over the graph states, representing the evolution of a population consisting of a continuum number of identical copies of a process modeled by the game graph. We consider synchronization objectives, which require the probability mass to accumulate in a set of target states, either always, once, infinitely often, or always after some point in the outcome sequence; and the winning modes of sure winning (if the accumulated probability is equal to 1) and almost-sure winning (if the accumulated probability is arbitrarily close to 1).
We present algorithms to compute the set of winning distributions for each of these synchronization modes, showing that the corresponding decision problem is PSPACE-complete for synchronizing once and infinitely often and PTIME-complete for synchronizing always and always after some point. These bounds are remarkably in line with the special case of MDPs, while the algorithmic solution and proof technique are considerably more involved, even for deterministic games. This is because those games have a flavor of imperfect information, in particular they are not determined and randomized strategies need to be considered, even if there is no stochastic choice in the game graph. Moreover, in combination with stochasticity in the game graph, finite-memory strategies are not sufficient in general.
{"title":"Stochastic Games with Synchronization Objectives","authors":"Laurent Doyen","doi":"https://dl.acm.org/doi/10.1145/3588866","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3588866","url":null,"abstract":"<p>We consider two-player stochastic games played on a finite graph for infinitely many rounds. Stochastic games generalize both Markov decision processes (MDP) by adding an adversary player, and two-player deterministic games by adding stochasticity. The outcome of the game is a sequence of distributions over the graph states, representing the evolution of a population consisting of a continuum number of identical copies of a process modeled by the game graph. We consider synchronization objectives, which require the probability mass to accumulate in a set of target states, either always, once, infinitely often, or always after some point in the outcome sequence; and the winning modes of sure winning (if the accumulated probability is equal to 1) and almost-sure winning (if the accumulated probability is arbitrarily close to 1).</p><p>We present algorithms to compute the set of winning distributions for each of these synchronization modes, showing that the corresponding decision problem is PSPACE-complete for synchronizing once and infinitely often and PTIME-complete for synchronizing always and always after some point. These bounds are remarkably in line with the special case of MDPs, while the algorithmic solution and proof technique are considerably more involved, even for deterministic games. This is because those games have a flavor of imperfect information, in particular they are not determined and randomized strategies need to be considered, even if there is no stochastic choice in the game graph. Moreover, in combination with stochasticity in the game graph, finite-memory strategies are not sufficient in general.</p>","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"45 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2023-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138525722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-24DOI: https://dl.acm.org/doi/10.1145/3588564
Moran Feldman, Ashkan Norouzi-Fard, Ola Svensson, Rico Zenklusen
We consider the classical problem of maximizing a monotone submodular function subject to a cardinality constraint, which, due to its numerous applications, has recently been studied in various computational models. We consider a clean multi-player model that lies between the offline and streaming model, and study it under the aspect of one-way communication complexity. Our model captures the streaming setting (by considering a large number of players), and, in addition, two player approximation results for it translate into the robust setting. We present tight one-way communication complexity results for our model, which, due to the above-mentioned connections, have multiple implications in the data stream and robust setting.
Even for just two players, a prior information-theoretic hardness result implies that no approximation factor above 1/2 can be achieved in our model, if only queries to feasible sets, i.e., sets respecting the cardinality constraint, are allowed. We show that the possibility of querying infeasible sets can actually be exploited to beat this bound, by presenting a tight 2/3-approximation taking exponential time, and an efficient 0.514-approximation. To the best of our knowledge, this is the first example where querying a submodular function on infeasible sets leads to provably better results. Through the above-mentioned link to the (non-streaming) robust setting, both of these algorithms improve on the current state-of-the-art for robust submodular maximization, showing that approximation factors beyond 1/2 are possible. Moreover, exploiting the link of our model to streaming, we settle the approximability for streaming algorithms by presenting a tight 1/2 + ε hardness result, based on the construction of a new family of coverage functions. This improves on a prior 0.586 hardness and matches, up to an arbitrarily small margin, the best known approximation algorithm.
{"title":"The One-Way Communication Complexity of Submodular Maximization with Applications to Streaming and Robustness","authors":"Moran Feldman, Ashkan Norouzi-Fard, Ola Svensson, Rico Zenklusen","doi":"https://dl.acm.org/doi/10.1145/3588564","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3588564","url":null,"abstract":"<p>We consider the classical problem of maximizing a monotone submodular function subject to a cardinality constraint, which, due to its numerous applications, has recently been studied in various computational models. We consider a clean multi-player model that lies between the offline and streaming model, and study it under the aspect of one-way communication complexity. Our model captures the streaming setting (by considering a large number of players), and, in addition, two player approximation results for it translate into the robust setting. We present tight one-way communication complexity results for our model, which, due to the above-mentioned connections, have multiple implications in the data stream and robust setting. </p><p>Even for just two players, a prior information-theoretic hardness result implies that no approximation factor above 1/2 can be achieved in our model, if only queries to feasible sets, i.e., sets respecting the cardinality constraint, are allowed. We show that the possibility of querying infeasible sets can actually be exploited to beat this bound, by presenting a tight 2/3-approximation taking exponential time, and an efficient 0.514-approximation. To the best of our knowledge, this is the first example where querying a submodular function on infeasible sets leads to provably better results. Through the above-mentioned link to the (non-streaming) robust setting, both of these algorithms improve on the current state-of-the-art for robust submodular maximization, showing that approximation factors beyond 1/2 are possible. Moreover, exploiting the link of our model to streaming, we settle the approximability for streaming algorithms by presenting a tight 1/2 + ε hardness result, based on the construction of a new family of coverage functions. This improves on a prior 0.586 hardness and matches, up to an arbitrarily small margin, the best known approximation algorithm.</p>","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"30 9-10","pages":""},"PeriodicalIF":2.5,"publicationDate":"2023-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138525716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Moran Feldman, A. Norouzi-Fard, O. Svensson, R. Zenklusen
We consider the classical problem of maximizing a monotone submodular function subject to a cardinality constraint, which, due to its numerous applications, has recently been studied in various computational models. We consider a clean multiplayer model that lies between the offline and streaming model, and study it under the aspect of one-way communication complexity. Our model captures the streaming setting (by considering a large number of players), and, in addition, two-player approximation results for it translate into the robust setting. We present tight one-way communication complexity results for our model, which, due to the connections mentioned previously, have multiple implications in the data stream and robust setting. Even for just two players, a prior information-theoretic hardness result implies that no approximation factor above 1/2 can be achieved in our model, if only queries to feasible sets (i.e., sets respecting the cardinality constraint) are allowed. We show that the possibility of querying infeasible sets can actually be exploited to beat this bound, by presenting a tight 2/3-approximation taking exponential time, and an efficient 0.514-approximation. To the best of our knowledge, this is the first example where querying a submodular function on infeasible sets leads to provably better results. Through the link to the (non-streaming) robust setting mentioned previously, both of these algorithms improve on the current state of the art for robust submodular maximization, showing that approximation factors beyond 1/2 are possible. Moreover, exploiting the link of our model to streaming, we settle the approximability for streaming algorithms by presenting a tight 1/2+ɛ hardness result, based on the construction of a new family of coverage functions. This improves on a prior 0.586 hardness and matches, up to an arbitrarily small margin, the best-known approximation algorithm.
{"title":"The One-Way Communication Complexity of Submodular Maximization with Applications to Streaming and Robustness","authors":"Moran Feldman, A. Norouzi-Fard, O. Svensson, R. Zenklusen","doi":"10.1145/3588564","DOIUrl":"https://doi.org/10.1145/3588564","url":null,"abstract":"We consider the classical problem of maximizing a monotone submodular function subject to a cardinality constraint, which, due to its numerous applications, has recently been studied in various computational models. We consider a clean multiplayer model that lies between the offline and streaming model, and study it under the aspect of one-way communication complexity. Our model captures the streaming setting (by considering a large number of players), and, in addition, two-player approximation results for it translate into the robust setting. We present tight one-way communication complexity results for our model, which, due to the connections mentioned previously, have multiple implications in the data stream and robust setting. Even for just two players, a prior information-theoretic hardness result implies that no approximation factor above 1/2 can be achieved in our model, if only queries to feasible sets (i.e., sets respecting the cardinality constraint) are allowed. We show that the possibility of querying infeasible sets can actually be exploited to beat this bound, by presenting a tight 2/3-approximation taking exponential time, and an efficient 0.514-approximation. To the best of our knowledge, this is the first example where querying a submodular function on infeasible sets leads to provably better results. Through the link to the (non-streaming) robust setting mentioned previously, both of these algorithms improve on the current state of the art for robust submodular maximization, showing that approximation factors beyond 1/2 are possible. Moreover, exploiting the link of our model to streaming, we settle the approximability for streaming algorithms by presenting a tight 1/2+ɛ hardness result, based on the construction of a new family of coverage functions. This improves on a prior 0.586 hardness and matches, up to an arbitrarily small margin, the best-known approximation algorithm.","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"115 1","pages":"1 - 52"},"PeriodicalIF":2.5,"publicationDate":"2023-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80286072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The Exponential-Time Hypothesis (ETH) is a strengthening of the 𝒫 ≠ 𝒩𝒫 conjecture, stating that 3- SAT on n variables cannot be solved in (uniform) time 2 εċ n , for some ε > 0. In recent years, analogous hypotheses that are “exponentially strong” forms of other classical complexity conjectures (such as 𝒩𝒫⊈ ℬ𝒫𝒫 or co 𝒩𝒫⊈𝒩𝒫) have also been introduced and have become widely influential. In this work, we focus on the interaction of exponential-time hypotheses with the fundamental and closely related questions of derandomization and circuit lower bounds . We show that even relatively mild variants of exponential-time hypotheses have far-reaching implications to derandomization, circuit lower bounds, and the connections between the two. Specifically, we prove that: (1) The Randomized Exponential-Time Hypothesis (rETH) implies that ℬ𝒫𝒫 can be simulated on “average-case” in deterministic (nearly-)polynomial-time (i.e., in time 2 Õ(log( n )) = n loglog( n ) O(1) ). The derandomization relies on a conditional construction of a pseudorandom generator with near-exponential stretch (i.e., with seed length Õ(log ( n ))); this significantly improves the state-of-the-art in uniform “hardness-to-randomness” results, which previously only yielded pseudorandom generators with sub-exponential stretch from such hypotheses. (2) The Non-Deterministic Exponential-Time Hypothesis (NETH) implies that derandomization of ℬ𝒫𝒫 is completely equivalent to circuit lower bounds against ℰ, and in particular that pseudorandom generators are necessary for derandomization. In fact, we show that the foregoing equivalence follows from a very weak version of NETH, and we also show that this very weak version is necessary to prove a slightly stronger conclusion that we deduce from it. Last, we show that disproving certain exponential-time hypotheses requires proving breakthrough circuit lower bounds. In particular, if CircuitSAT for circuits over n bits of size poly(n) can be solved by probabilistic algorithms in time 2 n /polylog(n) , then ℬ𝒫ℰ does not have circuits of quasilinear size.
{"title":"On Exponential-time Hypotheses, Derandomization, and Circuit Lower Bounds","authors":"Lijie Chen, Ron Rothblum, Roei Tell, Eylon Yogev","doi":"10.1145/3593581","DOIUrl":"https://doi.org/10.1145/3593581","url":null,"abstract":"The Exponential-Time Hypothesis (ETH) is a strengthening of the 𝒫 ≠ 𝒩𝒫 conjecture, stating that 3- SAT on n variables cannot be solved in (uniform) time 2 εċ n , for some ε > 0. In recent years, analogous hypotheses that are “exponentially strong” forms of other classical complexity conjectures (such as 𝒩𝒫⊈ ℬ𝒫𝒫 or co 𝒩𝒫⊈𝒩𝒫) have also been introduced and have become widely influential. In this work, we focus on the interaction of exponential-time hypotheses with the fundamental and closely related questions of derandomization and circuit lower bounds . We show that even relatively mild variants of exponential-time hypotheses have far-reaching implications to derandomization, circuit lower bounds, and the connections between the two. Specifically, we prove that: (1) The Randomized Exponential-Time Hypothesis (rETH) implies that ℬ𝒫𝒫 can be simulated on “average-case” in deterministic (nearly-)polynomial-time (i.e., in time 2 Õ(log( n )) = n loglog( n ) O(1) ). The derandomization relies on a conditional construction of a pseudorandom generator with near-exponential stretch (i.e., with seed length Õ(log ( n ))); this significantly improves the state-of-the-art in uniform “hardness-to-randomness” results, which previously only yielded pseudorandom generators with sub-exponential stretch from such hypotheses. (2) The Non-Deterministic Exponential-Time Hypothesis (NETH) implies that derandomization of ℬ𝒫𝒫 is completely equivalent to circuit lower bounds against ℰ, and in particular that pseudorandom generators are necessary for derandomization. In fact, we show that the foregoing equivalence follows from a very weak version of NETH, and we also show that this very weak version is necessary to prove a slightly stronger conclusion that we deduce from it. Last, we show that disproving certain exponential-time hypotheses requires proving breakthrough circuit lower bounds. In particular, if CircuitSAT for circuits over n bits of size poly(n) can be solved by probabilistic algorithms in time 2 n /polylog(n) , then ℬ𝒫ℰ does not have circuits of quasilinear size.","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"96 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135568383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-18DOI: https://dl.acm.org/doi/10.1145/3584699
Arik Rinberg, Idit Keidar
<p>Big data processing systems often employ batched updates and data sketches to estimate certain properties of large data. For example, a <i>CountMin sketch</i> approximates the frequencies at which elements occur in a data stream, and a <i>batched counter</i> counts events in batches. This article focuses on correctness criteria for concurrent implementations of such objects. Specifically, we consider <i>quantitative</i> objects, whose return values are from an ordered domain, with a particular emphasis on <i>(ε,δ)-bounded</i> objects that estimate a numerical quantity with an error of at most ε with probability at least 1 - δ.</p><p>The de facto correctness criterion for concurrent objects is linearizability. Intuitively, under linearizability, when a read overlaps an update, it must return the object’s value either before the update or after it. Consider, for example, a single batched increment operation that counts three new events, bumping a batched counter’s value from 7 to 10. In a linearizable implementation of the counter, a read overlapping this update must return either 7 or 10. We observe, however, that in typical use cases, any <i>intermediate</i> value between 7 and 10 would also be acceptable. To capture this additional degree of freedom, we propose <i>Intermediate Value Linearizability (IVL)</i>, a new correctness criterion that relaxes linearizability to allow returning intermediate values, for instance, 8 in the example above. Roughly speaking, IVL allows reads to return any value that is bounded between two return values that are legal under linearizability.</p><p>A key feature of IVL is that we can prove that concurrent IVL implementations of (ε,δ)-bounded objects are themselves (ε,δ)-bounded. To illustrate the power of this result, we give a straightforward and efficient concurrent implementation of an (ε,δ)-bounded CountMin sketch, which is IVL (albeit not linearizable).</p><p>We present four examples for IVL objects, each showcasing a different way of using IVL. The first is a simple wait-free IVL batched counter, with <i>O</i>(1) step complexity for update. The next considers an (ε,δ)-bounded CountMin sketch and further shows how to relax IVL using the notion of <i>r</i>-relaxation. Our third example is a non-atomic iterator over a data structure. In this example, we augment the data structure with an <i>auxiliary history variable</i> state that includes “tombstones” for items deleted from the data structure. Here, IVL semantics are required at the augmented level. Finally, using a <i>priority queue</i>, we show that some objects require IVL to be paired with other correctness criteria; indeed, a natural correctness notion for a concurrent priority queue is IVL coupled with sequential consistency.</p><p>Last, we show that IVL allows for inherently cheaper implementations than linearizable ones. In particular, we show a lower bound of Ω (<i>n</i>) on the step complexity of the update operation of any wait-free linearizable
大数据处理系统经常使用批量更新和数据草图来估计大数据的某些属性。例如,CountMin草图近似于元素在数据流中出现的频率,批处理计数器分批计数事件。本文主要关注这类对象的并发实现的正确性标准。具体地说,我们考虑定量对象,其返回值来自有序域,特别强调(ε,δ)有界对象,其估计数值数量的误差最多为ε,概率至少为1 - δ。并发对象的事实上的正确性标准是线性化。直观地说,在线性化条件下,当读取操作与更新操作重叠时,它必须返回对象的值,要么在更新操作之前,要么在更新操作之后。例如,考虑单个批处理增量操作,该操作计数三个新事件,将批处理计数器的值从7增加到10。在可线性化的计数器实现中,与此更新重叠的读操作必须返回7或10。然而,我们观察到,在典型的用例中,7到10之间的任何中间值也是可以接受的。为了获得这个额外的自由度,我们提出了中间值线性化(Intermediate Value Linearizability, IVL),这是一个新的正确性标准,它放宽了线性化,允许返回中间值,例如上面的例子中的8。粗略地说,IVL允许读取返回在线性化条件下合法的两个返回值之间的任何值。IVL的一个关键特征是我们可以证明(ε,δ)有界对象的并发IVL实现本身是(ε,δ)有界的。为了说明这个结果的力量,我们给出了一个(ε,δ)有界CountMin草图的简单有效的并发实现,它是IVL(尽管不是线性化的)。我们给出了IVL对象的四个示例,每个示例都展示了使用IVL的不同方式。第一个是简单的无等待IVL批处理计数器,更新的步骤复杂度为0(1)步。接下来考虑一个(ε,δ)有界的CountMin草图,并进一步展示如何使用r-松弛的概念来松弛IVL。第三个例子是数据结构上的非原子迭代器。在这个例子中,我们用一个辅助的历史变量状态来扩展数据结构,其中包括从数据结构中删除的项的“墓碑”。这里,在增强级别需要IVL语义。最后,使用优先级队列,我们展示了一些对象需要IVL与其他正确性标准配对;实际上,并发优先级队列的自然正确性概念是IVL与顺序一致性相结合。最后,我们展示了IVL允许比线性化实现更便宜的实现。特别是,我们展示了来自单写多读寄存器的任何无等待线性批处理计数器的更新操作的步复杂度的下界Ω (n),这比我们的O(1) IVL实现更昂贵。
{"title":"Intermediate Value Linearizability: A Quantitative Correctness Criterion","authors":"Arik Rinberg, Idit Keidar","doi":"https://dl.acm.org/doi/10.1145/3584699","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3584699","url":null,"abstract":"<p>Big data processing systems often employ batched updates and data sketches to estimate certain properties of large data. For example, a <i>CountMin sketch</i> approximates the frequencies at which elements occur in a data stream, and a <i>batched counter</i> counts events in batches. This article focuses on correctness criteria for concurrent implementations of such objects. Specifically, we consider <i>quantitative</i> objects, whose return values are from an ordered domain, with a particular emphasis on <i>(ε,δ)-bounded</i> objects that estimate a numerical quantity with an error of at most ε with probability at least 1 - δ.</p><p>The de facto correctness criterion for concurrent objects is linearizability. Intuitively, under linearizability, when a read overlaps an update, it must return the object’s value either before the update or after it. Consider, for example, a single batched increment operation that counts three new events, bumping a batched counter’s value from 7 to 10. In a linearizable implementation of the counter, a read overlapping this update must return either 7 or 10. We observe, however, that in typical use cases, any <i>intermediate</i> value between 7 and 10 would also be acceptable. To capture this additional degree of freedom, we propose <i>Intermediate Value Linearizability (IVL)</i>, a new correctness criterion that relaxes linearizability to allow returning intermediate values, for instance, 8 in the example above. Roughly speaking, IVL allows reads to return any value that is bounded between two return values that are legal under linearizability.</p><p>A key feature of IVL is that we can prove that concurrent IVL implementations of (ε,δ)-bounded objects are themselves (ε,δ)-bounded. To illustrate the power of this result, we give a straightforward and efficient concurrent implementation of an (ε,δ)-bounded CountMin sketch, which is IVL (albeit not linearizable).</p><p>We present four examples for IVL objects, each showcasing a different way of using IVL. The first is a simple wait-free IVL batched counter, with <i>O</i>(1) step complexity for update. The next considers an (ε,δ)-bounded CountMin sketch and further shows how to relax IVL using the notion of <i>r</i>-relaxation. Our third example is a non-atomic iterator over a data structure. In this example, we augment the data structure with an <i>auxiliary history variable</i> state that includes “tombstones” for items deleted from the data structure. Here, IVL semantics are required at the augmented level. Finally, using a <i>priority queue</i>, we show that some objects require IVL to be paired with other correctness criteria; indeed, a natural correctness notion for a concurrent priority queue is IVL coupled with sequential consistency.</p><p>Last, we show that IVL allows for inherently cheaper implementations than linearizable ones. In particular, we show a lower bound of Ω (<i>n</i>) on the step complexity of the update operation of any wait-free linearizable","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"43 11-12","pages":""},"PeriodicalIF":2.5,"publicationDate":"2023-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138525676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-18DOI: https://dl.acm.org/doi/10.1145/3578574
Peyman Afshani, Pingan Cheng
<p>In the semialgebraic range searching problem, we are given a set of <i>n</i> points in ℝ<sup><i>d</i></sup>, and we want to preprocess the points such that for any query range belonging to a family of constant complexity semialgebraic sets (Tarski cells), all the points intersecting the range can be reported or counted efficiently. When the ranges are composed of simplices, the problem is well-understood: It can be solved using <i>S(n)</i> space and with <i>Q(n)</i> query time with (S(n)Q(n)^d = tilde{O}(n^d),) where the (tilde{O}(cdot)) notation hides polylogarithmic factors and this trade-off is tight (up to <i>n</i><sup><i>o</i>(1)</sup> factors). In particular, there exist “low space” structures that use <i>O(n)</i> space with <i>O</i>(<i>n</i><sup>1-1/<i>d</i>}</sup>) query time [8, 25] and “fast query” structures that use <i>O</i>(<i>n</i><sup><i>d</i></sup>) space with <i>O</i>(log <i>n</i>) query time [9]. However, for general semialgebraic ranges, only “low space” solutions are known, but the best solutions [7] match the same trade-off curve as simplex queries, with <i>O</i>(<i>n</i>) space and (tilde{O}(n^{1-1/d})) query time. It has been conjectured that the same could be done for the “fast query” case, but this open problem has stayed unresolved.</p><p>Here, we disprove this conjecture. We give the first nontrivial lower bounds for semialgebraic range searching and other related problems. More precisely, we show that any data structure for reporting the points between two concentric circles, a problem that we call 2D annulus reporting, with <i>Q</i>(<i>n</i>) query time must use (S(n)=overset{scriptscriptstyle o}{Omega }(n^3/Q(n)^5)) space, where the (overset{scriptscriptstyle o}{Omega }(cdot)) notation hides (n^{o(1)}) factors, meaning, for (Q(n)=log ^{O(1)}n), (overset{scriptscriptstyle o}{Omega }(n^3)) space must be used. In addition, we study the problem of reporting the subset of input points in a polynomial slab defined by (lbrace (x,y)in mathbb {R}^2:P(x)le yle P(x)+wrbrace), where (P(x)=sum _{i=0}^Delta a_i x^i) is a univariate polynomial of degree Δ and (a_0, ldots , a_Delta , w) are given at the query time, a problem that we call polynomial slab reporting. For this, we show a space lower bound of (overset{scriptscriptstyle o}{Omega }(n^{Delta +1}/Q(n)^{(Delta +3)Delta /2})), which implies that for (Q(n)=log ^{O(1)}n), we must use (overset{scriptscriptstyle o}{Omega }(n^{Delta +1})) space. We also consider the dual semialgebraic stabbing problems of semialgebraic range searching and present lower bounds for them. In particular, we show that in linear space, any data structure that solves 2D annulus stabbing problems must use (Omega (n^{2/3})) query time. Note that this almost matches the upper bound obtained by lifting 2D annuli to 3D. Like semialgebraic range searching, we also present lower bounds for general polynomial slab stabbing problems. Again, our lower bounds are almost tight for linear size data structures in th
{"title":"Lower Bounds for Semialgebraic Range Searching and Stabbing Problems","authors":"Peyman Afshani, Pingan Cheng","doi":"https://dl.acm.org/doi/10.1145/3578574","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3578574","url":null,"abstract":"<p>In the semialgebraic range searching problem, we are given a set of <i>n</i> points in ℝ<sup><i>d</i></sup>, and we want to preprocess the points such that for any query range belonging to a family of constant complexity semialgebraic sets (Tarski cells), all the points intersecting the range can be reported or counted efficiently. When the ranges are composed of simplices, the problem is well-understood: It can be solved using <i>S(n)</i> space and with <i>Q(n)</i> query time with (S(n)Q(n)^d = tilde{O}(n^d),) where the (tilde{O}(cdot)) notation hides polylogarithmic factors and this trade-off is tight (up to <i>n</i><sup><i>o</i>(1)</sup> factors). In particular, there exist “low space” structures that use <i>O(n)</i> space with <i>O</i>(<i>n</i><sup>1-1/<i>d</i>}</sup>) query time [8, 25] and “fast query” structures that use <i>O</i>(<i>n</i><sup><i>d</i></sup>) space with <i>O</i>(log <i>n</i>) query time [9]. However, for general semialgebraic ranges, only “low space” solutions are known, but the best solutions [7] match the same trade-off curve as simplex queries, with <i>O</i>(<i>n</i>) space and (tilde{O}(n^{1-1/d})) query time. It has been conjectured that the same could be done for the “fast query” case, but this open problem has stayed unresolved.</p><p>Here, we disprove this conjecture. We give the first nontrivial lower bounds for semialgebraic range searching and other related problems. More precisely, we show that any data structure for reporting the points between two concentric circles, a problem that we call 2D annulus reporting, with <i>Q</i>(<i>n</i>) query time must use (S(n)=overset{scriptscriptstyle o}{Omega }(n^3/Q(n)^5)) space, where the (overset{scriptscriptstyle o}{Omega }(cdot)) notation hides (n^{o(1)}) factors, meaning, for (Q(n)=log ^{O(1)}n), (overset{scriptscriptstyle o}{Omega }(n^3)) space must be used. In addition, we study the problem of reporting the subset of input points in a polynomial slab defined by (lbrace (x,y)in mathbb {R}^2:P(x)le yle P(x)+wrbrace), where (P(x)=sum _{i=0}^Delta a_i x^i) is a univariate polynomial of degree Δ and (a_0, ldots , a_Delta , w) are given at the query time, a problem that we call polynomial slab reporting. For this, we show a space lower bound of (overset{scriptscriptstyle o}{Omega }(n^{Delta +1}/Q(n)^{(Delta +3)Delta /2})), which implies that for (Q(n)=log ^{O(1)}n), we must use (overset{scriptscriptstyle o}{Omega }(n^{Delta +1})) space. We also consider the dual semialgebraic stabbing problems of semialgebraic range searching and present lower bounds for them. In particular, we show that in linear space, any data structure that solves 2D annulus stabbing problems must use (Omega (n^{2/3})) query time. Note that this almost matches the upper bound obtained by lifting 2D annuli to 3D. Like semialgebraic range searching, we also present lower bounds for general polynomial slab stabbing problems. Again, our lower bounds are almost tight for linear size data structures in th","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"45 9","pages":""},"PeriodicalIF":2.5,"publicationDate":"2023-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138525668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-03-25DOI: https://dl.acm.org/doi/10.1145/3578579
Ivan Geffner, Joseph Y. Halpern
Abraham, Dolev, Geffner, and Halpern [1] proved that, in asynchronous systems, a (k, t)-robust equilibrium for n players and a trusted mediator can be implemented without the mediator as long as n > 4(k+t), where an equilibrium is (k, t)-robust if, roughly speaking, no coalition of t players can decrease the payoff of any of the other players, and no coalition of k players can increase their payoff by deviating. We prove that this bound is tight, in the sense that if n ≤ 4(k+t) there exist (k, t)-robust equilibria with a mediator that cannot be implemented by the players alone. Even though implementing (k, t)-robust mediators seems closely related to implementing asynchronous multiparty (k+t)-secure computation [6], to the best of our knowledge there is no known straightforward reduction from one problem to another. Nevertheless, we show that there is a non-trivial reduction from a slightly weaker notion of (k+t)-secure computation, which we call (k+t)-strict secure computation, to implementing (k, t)-robust mediators. We prove the desired lower bound by showing that there are functions on n variables that cannot be (k+t)-strictly securely computed if n ≤ 4(k+t). This also provides a simple alternative proof for the well-known lower bound of 4t+1 on asynchronous secure computation in the presence of up to t malicious agents [4, 8, 10].
Abraham, Dolev, Geffner, and Halpern[1]证明了在异步系统中,只要n >4(k+t),其中均衡是(k, t)-鲁棒性,粗略地说,如果t个参与者的联盟不能减少其他参与者的收益,并且k个参与者的联盟不能通过偏离来增加他们的收益。我们证明了这个界是紧的,即如果n≤4(k+t),存在(k, t)个具有中介的鲁棒均衡,且不能由参与人单独实现。尽管实现(k, t)健壮的中介器似乎与实现异步多方(k+t)安全计算密切相关[6],但据我们所知,没有已知的从一个问题到另一个问题的直接简化。然而,我们证明了从稍微弱一点的(k+t)安全计算的概念(我们称之为(k+t)严格安全计算)到实现(k, t)鲁棒中介的非平凡简化。我们通过证明n个变量上的函数不能是(k+t)——当n≤4(k+t)时严格安全计算,证明了期望的下界。这也为异步安全计算中存在多达t个恶意代理时众所周知的4t+1下界提供了一个简单的替代证明[4,8,10]。
{"title":"Lower Bounds on Implementing Mediators in Asynchronous Systems with Rational and Malicious Agents","authors":"Ivan Geffner, Joseph Y. Halpern","doi":"https://dl.acm.org/doi/10.1145/3578579","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3578579","url":null,"abstract":"<p>Abraham, Dolev, Geffner, and Halpern [1] proved that, in asynchronous systems, a <i>(k, t)-robust equilibrium</i> for <i>n</i> players and a trusted mediator can be implemented without the mediator as long as <i>n</i> > 4(<i>k+t</i>), where an equilibrium is (<i>k, t</i>)-robust if, roughly speaking, no coalition of <i>t</i> players can decrease the payoff of any of the other players, and no coalition of <i>k</i> players can increase their payoff by deviating. We prove that this bound is tight, in the sense that if <i>n</i> ≤ 4(<i>k+t</i>) there exist (<i>k, t</i>)-robust equilibria with a mediator that cannot be implemented by the players alone. Even though implementing (<i>k, t</i>)-robust mediators seems closely related to implementing asynchronous multiparty (<i>k+t</i>)-secure computation [6], to the best of our knowledge there is no known straightforward reduction from one problem to another. Nevertheless, we show that there is a non-trivial reduction from a slightly weaker notion of (<i>k+t</i>)-secure computation, which we call <i>(<i>k+t</i>)-strict secure computation</i>, to implementing (<i>k, t</i>)-robust mediators. We prove the desired lower bound by showing that there are functions on <i>n</i> variables that cannot be (<i>k+t</i>)-strictly securely computed if <i>n</i> ≤ 4(<i>k+t</i>). This also provides a simple alternative proof for the well-known lower bound of 4<i>t</i>+1 on asynchronous secure computation in the presence of up to <i>t</i> malicious agents [4, 8, 10].</p>","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"19 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2023-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138525666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-03-25DOI: https://dl.acm.org/doi/10.1145/3572918
Moritz Lichter
In the search for a logic capturing polynomial time the most promising candidates are Choiceless Polynomial Time (CPT) and rank logic. Rank logic extends fixed-point logic with counting by a rank operator over prime fields. We show that the isomorphism problem for CFI graphs over ℤ2i cannot be defined in rank logic, even if the base graph is totally ordered. However, CPT can define this isomorphism problem. We thereby separate rank logic from CPT and in particular from polynomial time.
{"title":"Separating Rank Logic from Polynomial Time","authors":"Moritz Lichter","doi":"https://dl.acm.org/doi/10.1145/3572918","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3572918","url":null,"abstract":"<p>In the search for a logic capturing polynomial time the most promising candidates are Choiceless Polynomial Time (CPT) and rank logic. Rank logic extends fixed-point logic with counting by a rank operator over prime fields. We show that the isomorphism problem for CFI graphs over ℤ<sub>2<sup><i>i</i></sup></sub> cannot be defined in rank logic, even if the base graph is totally ordered. However, CPT can define this isomorphism problem. We thereby separate rank logic from CPT and in particular from polynomial time.</p>","PeriodicalId":50022,"journal":{"name":"Journal of the ACM","volume":"32 7-8","pages":""},"PeriodicalIF":2.5,"publicationDate":"2023-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138525735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}