We use the reconfiguration framework to analyze problems that involve the rearrangement of items among groups. In various applications, a group of items could correspond to the files or jobs assigned to a particular machine, and the goal of rearrangement could be improving efficiency or increasing locality. To cover problems arising in a wide range of application areas, we define the general Repacking problem as the rearrangement of multisets of multisets. We present hardness results for the general case and algorithms for various restricted classes of instances. By limiting the total size of items in each multiset, our results can be viewed as an offline approach to Bin Packing, in which each bin is represented as a multiset. In addition to providing the first results on reconfiguration of multisets, our contributions open up several research avenues: the interplay between reconfiguration and online algorithms and parallel algorithms; the use of the tools of linear programming in reconfiguration; and, in the longer term, a focus on extra resources in reconfiguration. A preliminary version of this paper appeared in the proceedings of the 18th International Conference and Workshops on Algorithms and Computation (WALCOM 2024).
{"title":"Reconfiguration of Multisets with Applications to Bin Packing","authors":"Jeffrey Kam, Shahin Kamali, Avery Miller, Naomi Nishimura","doi":"10.1007/s00453-025-01324-w","DOIUrl":"10.1007/s00453-025-01324-w","url":null,"abstract":"<div><p>We use the reconfiguration framework to analyze problems that involve the rearrangement of items among groups. In various applications, a group of items could correspond to the files or jobs assigned to a particular machine, and the goal of rearrangement could be improving efficiency or increasing locality. To cover problems arising in a wide range of application areas, we define the general <span>Repacking</span> problem as the rearrangement of multisets of multisets. We present hardness results for the general case and algorithms for various restricted classes of instances. By limiting the total size of items in each multiset, our results can be viewed as an offline approach to <span>Bin Packing</span>, in which each bin is represented as a multiset. In addition to providing the first results on reconfiguration of multisets, our contributions open up several research avenues: the interplay between reconfiguration and online algorithms and parallel algorithms; the use of the tools of linear programming in reconfiguration; and, in the longer term, a focus on extra resources in reconfiguration. A preliminary version of this paper appeared in the proceedings of the 18th International Conference and Workshops on Algorithms and Computation (WALCOM 2024).</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 12","pages":"1933 - 1996"},"PeriodicalIF":0.7,"publicationDate":"2025-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145227882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-20DOI: 10.1007/s00453-025-01332-w
Sam Coy, Artur Czumaj, Gopinath Mishra, Anish Mukherjee
We consider two natural variants of the problem of minimum spanning tree ((text {MST})) of a graph in the parallel setting: MST verification (verifying if a given tree is an (text {MST})) and the sensitivity analysis of an MST (finding the lowest cost replacement edge for each edge of the (text {MST})). These two problems have been studied extensively for sequential algorithms and for parallel algorithms in the (textrm{PRAM}) model of computation. In this paper, we extend the study to the standard model of Massive Parallel Computation ((textrm{MPC})). It is known that for graphs of diameter D, the connectivity problem can be solved in (O(log D + log log n)) rounds on an (textrm{MPC}) with low local memory (each machine can store only (O(n^{delta })) words for an arbitrary constant (delta > 0)) and with linear global memory, that is, with optimal utilization. However, for the related task of finding an (text {MST}), we need (Omega (log D_{text {MST}})) rounds, where (D_{text {MST}}) denotes the diameter of the minimum spanning tree. The state of the art upper bound for (text {MST}) is (O(log n)) rounds; the result follows by simulating existing (textrm{PRAM}) algorithms. While this bound may be optimal for general graphs, the benchmark of connectivity and lower bound for (text {MST}) suggest the target bound of (O(log D_text {MST})) rounds, or possibly (O(log D_text {MST} + log log n)) rounds. As for now, we do not know if this bound is achievable for the (text {MST}) problem on an (textrm{MPC}) with low local memory and linear global memory. In this paper, we show that two natural variants of the (text {MST}) problem: (text {MST}) verification and sensitivity analysis of an (text {MST}), can be completed in (O(log D_T)) rounds on an (textrm{MPC}) with low local memory and with linear global memory, that is, with optimal utilization; here (D_T) is the diameter of the input “candidate (text {MST}) ” T. The algorithms asymptotically match our lower bound, conditioned on the 1-vs-2-cycle conjecture.
{"title":"Log-Diameter MST Verification and Sensitivity in MPC","authors":"Sam Coy, Artur Czumaj, Gopinath Mishra, Anish Mukherjee","doi":"10.1007/s00453-025-01332-w","DOIUrl":"10.1007/s00453-025-01332-w","url":null,"abstract":"<div><p>We consider two natural variants of the problem of minimum spanning tree (<span>(text {MST})</span>) of a graph in the parallel setting: <i>MST verification</i> (verifying if a given tree is an <span>(text {MST})</span>) and the <i>sensitivity analysis of an MST</i> (finding the lowest cost replacement edge for each edge of the <span>(text {MST})</span>). These two problems have been studied extensively for sequential algorithms and for parallel algorithms in the <span>(textrm{PRAM})</span> model of computation. In this paper, we extend the study to the standard model of <i>Massive Parallel Computation</i> (<span>(textrm{MPC})</span>). It is known that for graphs of diameter <i>D</i>, the connectivity problem can be solved in <span>(O(log D + log log n))</span> rounds on an <span>(textrm{MPC})</span> with <i>low local memory</i> (each machine can store only <span>(O(n^{delta }))</span> words for an arbitrary constant <span>(delta > 0)</span>) and with <i>linear global memory</i>, that is, with <i>optimal utilization</i>. However, for the related task of finding an <span>(text {MST})</span>, we need <span>(Omega (log D_{text {MST}}))</span> rounds, where <span>(D_{text {MST}})</span> denotes the diameter of the minimum spanning tree. The state of the art upper bound for <span>(text {MST})</span> is <span>(O(log n))</span> rounds; the result follows by simulating existing <span>(textrm{PRAM})</span> algorithms. While this bound may be optimal for general graphs, the benchmark of connectivity and lower bound for <span>(text {MST})</span> suggest the target bound of <span>(O(log D_text {MST}))</span> rounds, or possibly <span>(O(log D_text {MST} + log log n))</span> rounds. As for now, we do not know if this bound is achievable for the <span>(text {MST})</span> problem on an <span>(textrm{MPC})</span> with low local memory and linear global memory. In this paper, we show that two natural variants of the <span>(text {MST})</span> problem: <span>(text {MST})</span> verification and sensitivity analysis of an <span>(text {MST})</span>, can be completed in <span>(O(log D_T))</span> rounds on an <span>(textrm{MPC})</span> with low local memory and with linear global memory, that is, with optimal utilization; here <span>(D_T)</span> is the diameter of the input “candidate <span>(text {MST})</span> ” <i>T</i>. The algorithms asymptotically match our lower bound, conditioned on the 1-vs-2-cycle conjecture.</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 12","pages":"1899 - 1932"},"PeriodicalIF":0.7,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145227911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-19DOI: 10.1007/s00453-025-01329-5
N. R. Aravind, Roopam Saxena
In Path Set Packing, the input is an undirected graph G, a collection (mathcal{P}) of simple paths in G, and a positive integer k. The problem is to decide whether there exist k edge-disjoint paths in (mathcal{P}). We study the parameterized complexity of Path Set Packing with respect to both natural and structural parameters. We show that the problem is W[1]-hard with respect to vertex cover number, and W[1]-hard respect to pathwidth plus solution size when input graph is a grid. These results answer an open question raised in Xu and Zhang (in: Wang L, Zhu D (eds) Computing and combinatorics—24th international conference, COCOON 2018, Qing Dao, China, July 2–4, 2018, proceedings. Lecture notes in computer science, vol 10976, pp 305–315. Springer, 2018, https://doi.org/10.1007/978-3-319-94776-1_26). On the positive side, we present an FPT algorithm parameterized by feedback vertex number plus maximum degree, and present an FPT algorithm parameterized by treewidth plus maximum degree plus maximum length of a path in (mathcal{P}). These positive results complement the hardness of Path Set Packing with respect to any subset of the parameters used in the FPT algorithms. We also give a 4-approximation algorithm for maximum path set packing problem which runs in FPT time when parameterized by feedback edge number.
在路径集填充中,输入是一个无向图G, G中简单路径的集合(mathcal{P})和一个正整数k,问题是确定(mathcal{P})中是否存在k条不相交的路径。本文从自然参数和结构参数两方面研究了路径集填充的参数化复杂度。我们表明,当输入图是网格时,问题是W[1]-关于顶点覆盖数的困难,W[1]-关于路径宽度加解大小的困难。这些结果回答了Xu和Zhang提出的一个开放性问题(in: Wang L, Zhu D(编))计算与组合-第24届国际会议,COCOON 2018,中国青岛,2018年7月2-4日,proceedings。《计算机科学》,第10976卷,第305-315页。b施普林格,2018,https://doi.org/10.1007/978-3-319-94776-1_26)。在积极的方面,我们提出了一个参数化的FPT算法的反馈顶点数加上最大度,并提出了一个参数化的FPT算法的树宽加上最大度加上路径的最大长度(mathcal{P})。这些积极的结果补充了路径集填充相对于FPT算法中使用的任何参数子集的硬度。对于用反馈边数参数化的FPT时间内运行的最大路径集布局问题,给出了一个4逼近算法。
{"title":"Parameterized Complexity of Path Set Packing","authors":"N. R. Aravind, Roopam Saxena","doi":"10.1007/s00453-025-01329-5","DOIUrl":"10.1007/s00453-025-01329-5","url":null,"abstract":"<div><p>In <span>Path Set Packing</span>, the input is an undirected graph <i>G</i>, a collection <span>(mathcal{P})</span> of simple paths in <i>G</i>, and a positive integer <i>k</i>. The problem is to decide whether there exist <i>k</i> edge-disjoint paths in <span>(mathcal{P})</span>. We study the parameterized complexity of <span>Path Set Packing</span> with respect to both natural and structural parameters. We show that the problem is W[1]-hard with respect to vertex cover number, and W[1]-hard respect to pathwidth plus solution size when input graph is a grid. These results answer an open question raised in Xu and Zhang (in: Wang L, Zhu D (eds) Computing and combinatorics—24th international conference, COCOON 2018, Qing Dao, China, July 2–4, 2018, proceedings. Lecture notes in computer science, vol 10976, pp 305–315. Springer, 2018, https://doi.org/10.1007/978-3-319-94776-1_26). On the positive side, we present an FPT algorithm parameterized by feedback vertex number plus maximum degree, and present an FPT algorithm parameterized by treewidth plus maximum degree plus maximum length of a path in <span>(mathcal{P})</span>. These positive results complement the hardness of <span>Path Set Packing</span> with respect to any subset of the parameters used in the FPT algorithms. We also give a 4-approximation algorithm for maximum path set packing problem which runs in FPT time when parameterized by feedback edge number.</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 12","pages":"1864 - 1898"},"PeriodicalIF":0.7,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145227913","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-18DOI: 10.1007/s00453-025-01330-y
Joost Jorritsma, Johannes Lengler, Dirk Sudholt
Evolutionary algorithms (EAs) are general-purpose optimisation algorithms that maintain a population (multiset) of candidate solutions and apply variation operators to create new solutions called offspring. A new population is typically formed using one of two strategies: a ((mu +lambda )) EA (plus selection) keeps the best (mu ) search points out of the union of (mu ) parents in the old population and (lambda ) offspring, whereas a ((mu ,lambda )) EA (comma selection) discards all parents and only keeps the best (mu ) out of (lambda ) offspring. Comma selection may help to escape from local optima, however when and how it is beneficial is subject to an ongoing debate. We propose a new benchmark function to investigate the benefits of comma selection: the well known benchmark function OneMaxwith randomly planted local optima, generated by frozen noise. We show that comma selection (the ({(1,lambda )}) EA) is faster than plus selection (the ({(1+lambda )}) EA) on this benchmark, in a fixed-target scenario, and for offspring population sizes (lambda ) for which both algorithms behave differently. For certain parameters, the ({(1,lambda )}) EAfinds the target in (Theta (n ln n)) evaluations, with high probability (w.h.p.), while the ({(1+lambda )}) EAw.h.p. requires (omega (n^2)) evaluations. We further show that the advantage of comma selection is not arbitrarily large: w.h.p. comma selection outperforms plus selection at most by a factor of (O(n ln n)) for most reasonable parameter choices. We develop novel methods for analysing frozen noise and give powerful and general fixed-target results with tail bounds that are of independent interest.
{"title":"Comma Selection Outperforms Plus Selection on OneMax with Randomly Planted Optima","authors":"Joost Jorritsma, Johannes Lengler, Dirk Sudholt","doi":"10.1007/s00453-025-01330-y","DOIUrl":"10.1007/s00453-025-01330-y","url":null,"abstract":"<div><p>Evolutionary algorithms (EAs) are general-purpose optimisation algorithms that maintain a population (multiset) of candidate solutions and apply variation operators to create new solutions called offspring. A new population is typically formed using one of two strategies: a <span>((mu +lambda ))</span> EA (plus selection) keeps the best <span>(mu )</span> search points out of the union of <span>(mu )</span> parents in the old population and <span>(lambda )</span> offspring, whereas a <span>((mu ,lambda ))</span> EA (comma selection) discards all parents and only keeps the best <span>(mu )</span> out of <span>(lambda )</span> offspring. Comma selection may help to escape from local optima, however when and how it is beneficial is subject to an ongoing debate. We propose a new benchmark function to investigate the benefits of comma selection: the well known benchmark function <span>OneMax</span>with randomly planted local optima, generated by frozen noise. We show that comma selection (the <span>({(1,lambda )})</span> EA) is faster than plus selection (the <span>({(1+lambda )})</span> EA) on this benchmark, in a fixed-target scenario, and for offspring population sizes <span>(lambda )</span> for which both algorithms behave differently. For certain parameters, the <span>({(1,lambda )})</span> EAfinds the target in <span>(Theta (n ln n))</span> evaluations, with high probability (w.h.p.), while the <span>({(1+lambda )})</span> EAw.h.p. requires <span>(omega (n^2))</span> evaluations. We further show that the advantage of comma selection is not arbitrarily large: w.h.p. comma selection outperforms plus selection at most by a factor of <span>(O(n ln n))</span> for most reasonable parameter choices. We develop novel methods for analysing frozen noise and give powerful and general fixed-target results with tail bounds that are of independent interest.</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 12","pages":"1804 - 1863"},"PeriodicalIF":0.7,"publicationDate":"2025-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s00453-025-01330-y.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145227910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-13DOI: 10.1007/s00453-025-01339-3
Minglong Qin, Penghui Yao
This paper considers the decidability of fully quantum nonlocal games with noisy maximally entangled states. Fully quantum nonlocal games are a generalization of nonlocal games, where both questions and answers are quantum and the referee performs a binary POVM measurement to decide whether they win the game after receiving the quantum answers from the players. The quantum value of a fully quantum nonlocal game is the supremum of the probability that they win the game, where the supremum is taken over all the possible entangled states shared between the players and all the valid quantum operations performed by the players. The seminal work (text {MIP}^*=text {RE}) ( Ji et al. MIP ∗ = RE, 2020; Ji et al. Quantum soundness of the classical low individual degree test, 2020) implies that it is undecidable to approximate the quantum value of a fully nonlocal game. This still holds even if the players are only allowed to share (arbitrarily many copies of) maximally entangled states. This paper investigates the case that the shared maximally entangled states are noisy. We prove that there is a computable upper bound on the copies of noisy maximally entangled states for the players to win a fully quantum nonlocal game with a probability arbitrarily close to the quantum value. This implies that it is decidable to approximate the quantum values of these games. Hence, the hardness of approximating the quantum value of a fully quantum nonlocal game is not robust against the noise in the shared states. This paper is built on the framework for the decidability of non-interactive simulations of joint distributions (Ghazi et al. in: 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS), Los Alamitos, 2016; De et al. in: Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms, Philadelphia, 2018; Ghazi et al. Proceedings of the 33rd Computational Complexity Conference, 2018) and generalizes the analogous result for nonlocal games in Qin and Yao (SIAM J Comput 50(6):1800–1891, 2021). We extend the theory of Fourier analysis to the space of super-operators and prove several key results including an invariance principle and a dimension reduction for super-operators. These results are interesting in their own right and are believed to have further applications.
研究了具有噪声最大纠缠态的全量子非局部对策的可判决性。全量子非局部博弈是非局部博弈的泛化,其中问题和答案都是量子的,裁判在收到玩家的量子答案后执行二进制POVM测量来决定他们是否赢得比赛。全量子非局域博弈的量子值是他们赢得博弈的概率的最高值,其中最高值是参与者之间共享的所有可能的纠缠态和参与者执行的所有有效量子操作。开创性的工作(text {MIP}^*=text {RE}) (Ji et al.)Mip∗= re, 2020;Ji等人。经典的低个体度检验(2020)的量子稳健性意味着完全非局部博弈的量子值近似是不可确定的。即使玩家只被允许共享(任意多个副本)最大纠缠状态,这一点仍然成立。本文研究了共享最大纠缠态是有噪声的情况。我们证明了在一个概率任意接近于量子值的全量子非局部博弈中,参与者在噪声最大纠缠态的副本上存在一个可计算的上界。这意味着这些游戏的量子值是可以确定的。因此,近似全量子非局部对策的量子值的硬度对共享状态中的噪声不具有鲁棒性。本文建立在联合分布的非交互式模拟的可确定性框架(Ghazi等人:2016年IEEE第57届计算机科学基础年度研讨会(FOCS), Los Alamitos, 2016;De等人:第29届ACM-SIAM离散算法研讨会论文集,费城,2018;Ghazi等人。并对秦尧非局部博弈的模拟结果进行了推广[J] .计算学报,50(6):1800-1891,2021)。将傅里叶分析理论推广到超级算子空间,证明了超级算子的不变性原理和降维性。这些结果本身就很有趣,并被认为有进一步的应用。
{"title":"Decidability of Fully Quantum Nonlocal Games with Noisy Maximally Entangled States","authors":"Minglong Qin, Penghui Yao","doi":"10.1007/s00453-025-01339-3","DOIUrl":"10.1007/s00453-025-01339-3","url":null,"abstract":"<div><p>This paper considers the decidability of fully quantum nonlocal games with noisy maximally entangled states. Fully quantum nonlocal games are a generalization of nonlocal games, where both questions and answers are quantum and the referee performs a binary POVM measurement to decide whether they win the game after receiving the quantum answers from the players. The quantum value of a fully quantum nonlocal game is the supremum of the probability that they win the game, where the supremum is taken over all the possible entangled states shared between the players and all the valid quantum operations performed by the players. The seminal work <span>(text {MIP}^*=text {RE})</span> ( Ji et al. MIP ∗ = RE, 2020; Ji et al. Quantum soundness of the classical low individual degree test, 2020) implies that it is undecidable to approximate the quantum value of a fully nonlocal game. This still holds even if the players are only allowed to share (arbitrarily many copies of) maximally entangled states. This paper investigates the case that the shared maximally entangled states are noisy. We prove that there is a computable upper bound on the copies of noisy maximally entangled states for the players to win a fully quantum nonlocal game with a probability arbitrarily close to the quantum value. This implies that it is decidable to approximate the quantum values of these games. Hence, the hardness of approximating the quantum value of a fully quantum nonlocal game is not robust against the noise in the shared states. This paper is built on the framework for the decidability of non-interactive simulations of joint distributions (Ghazi et al. in: 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS), Los Alamitos, 2016; De et al. in: Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms, Philadelphia, 2018; Ghazi et al. Proceedings of the 33rd Computational Complexity Conference, 2018) and generalizes the analogous result for nonlocal games in Qin and Yao (SIAM J Comput 50(6):1800–1891, 2021). We extend the theory of Fourier analysis to the space of super-operators and prove several key results including an invariance principle and a dimension reduction for super-operators. These results are interesting in their own right and are believed to have further applications.</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 12","pages":"1732 - 1803"},"PeriodicalIF":0.7,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145227914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-13DOI: 10.1007/s00453-025-01338-4
Karthik C.S., Euiwoong Lee, Pasin Manurangsi
Parameterized Inapproximability Hypothesis ((textsf{PIH})) is a central question in the field of parameterized complexity. (textsf{PIH}) asserts that given as input a 2-(textsf{CSP}) on k variables and alphabet size n, it is (textsf{W})[1]-hard parameterized by k to distinguish if the input is perfectly satisfiable or if every assignment to the input violates 1% of the constraints. An important implication of (textsf{PIH}) is that it yields the tight parameterized inapproximability of the (k)-(textsf{maxcoverage}) problem. In the (k)-(textsf{maxcoverage}) problem, we are given as input a set system, a threshold (tau >0), and a parameter k and the goal is to determine if there exist k sets in the input whose union is at least (tau ) fraction of the entire universe. (textsf{PIH}) is known to imply that it is (textsf{W})[1]-hard parameterized by k to distinguish if there are k input sets whose union is at least (tau ) fraction of the universe or if the union of every k input sets is not much larger than (tau cdot (1-frac{1}{e})) fraction of the universe. In this work we present a gap preserving (textsf{FPT}) reduction (in the reverse direction) from the (k)-(textsf{maxcoverage}) problem to the aforementioned 2-(textsf{CSP}) problem, thus showing that the assertion that approximating the (k)-(textsf{maxcoverage}) problem to some constant factor is (textsf{W})[1]-hard implies (textsf{PIH}). In addition, we present a gap preserving (textsf{FPT}) reduction from the (k)-(textsf{median}) problem (in general metrics) to the (k)-(textsf{maxcoverage}) problem, further highlighting the power of gap preserving (textsf{FPT}) reductions over classical gap preserving polynomial time reductions.
参数化不可逼近性假说((textsf{PIH}))是参数化复杂性研究领域的一个核心问题。(textsf{PIH})断言,给定k个变量和字母大小为n的输入为2- (textsf{CSP}),则(textsf{W})[1]-很难用k参数化,以区分输入是完全可满足的,还是对输入的每个赋值都违反1% of the constraints. An important implication of (textsf{PIH}) is that it yields the tight parameterized inapproximability of the (k)-(textsf{maxcoverage}) problem. In the (k)-(textsf{maxcoverage}) problem, we are given as input a set system, a threshold (tau >0), and a parameter k and the goal is to determine if there exist k sets in the input whose union is at least (tau ) fraction of the entire universe. (textsf{PIH}) is known to imply that it is (textsf{W})[1]-hard parameterized by k to distinguish if there are k input sets whose union is at least (tau ) fraction of the universe or if the union of every k input sets is not much larger than (tau cdot (1-frac{1}{e})) fraction of the universe. In this work we present a gap preserving (textsf{FPT}) reduction (in the reverse direction) from the (k)-(textsf{maxcoverage}) problem to the aforementioned 2-(textsf{CSP}) problem, thus showing that the assertion that approximating the (k)-(textsf{maxcoverage}) problem to some constant factor is (textsf{W})[1]-hard implies (textsf{PIH}). In addition, we present a gap preserving (textsf{FPT}) reduction from the (k)-(textsf{median}) problem (in general metrics) to the (k)-(textsf{maxcoverage}) problem, further highlighting the power of gap preserving (textsf{FPT}) reductions over classical gap preserving polynomial time reductions.
{"title":"On Equivalence of Parameterized Inapproximability of k-Median, k-Max-Coverage, and 2-CSP","authors":"Karthik C.S., Euiwoong Lee, Pasin Manurangsi","doi":"10.1007/s00453-025-01338-4","DOIUrl":"10.1007/s00453-025-01338-4","url":null,"abstract":"<div><p>Parameterized Inapproximability Hypothesis (<span>(textsf{PIH})</span>) is a central question in the field of parameterized complexity. <span>(textsf{PIH})</span> asserts that given as input a 2-<span>(textsf{CSP})</span> on <i>k</i> variables and alphabet size <i>n</i>, it is <span>(textsf{W})</span>[1]-hard parameterized by <i>k</i> to distinguish if the input is perfectly satisfiable or if every assignment to the input violates 1% of the constraints. An important implication of <span>(textsf{PIH})</span> is that it yields the tight parameterized inapproximability of the <span>(k)</span>-<span>(textsf{maxcoverage})</span> problem. In the <span>(k)</span>-<span>(textsf{maxcoverage})</span> problem, we are given as input a set system, a threshold <span>(tau >0)</span>, and a parameter <i>k</i> and the goal is to determine if there exist <i>k</i> sets in the input whose union is at least <span>(tau )</span> fraction of the entire universe. <span>(textsf{PIH})</span> is known to imply that it is <span>(textsf{W})</span>[1]-hard parameterized by <i>k</i> to distinguish if there are <i>k</i> input sets whose union is at least <span>(tau )</span> fraction of the universe or if the union of every <i>k</i> input sets is not much larger than <span>(tau cdot (1-frac{1}{e}))</span> fraction of the universe. In this work we present a gap preserving <span>(textsf{FPT})</span> reduction (in the reverse direction) from the <span>(k)</span>-<span>(textsf{maxcoverage})</span> problem to the aforementioned 2-<span>(textsf{CSP})</span> problem, thus showing that the assertion that approximating the <span>(k)</span>-<span>(textsf{maxcoverage})</span> problem to some constant factor is <span>(textsf{W})</span>[1]-hard implies <span>(textsf{PIH})</span>. In addition, we present a gap preserving <span>(textsf{FPT})</span> reduction from the <span>(k)</span>-<span>(textsf{median})</span> problem (in general metrics) to the <span>(k)</span>-<span>(textsf{maxcoverage})</span> problem, further highlighting the power of gap preserving <span>(textsf{FPT})</span> reductions over classical gap preserving polynomial time reductions.</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 12","pages":"1711 - 1731"},"PeriodicalIF":0.7,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145227912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-07DOI: 10.1007/s00453-025-01336-6
Václav Blažej, Satyabrata Jana, M. S. Ramanujan, Peter Strulo
In this paper, we study the Eulerian Strong Component Arc Deletion problem, where the input is a directed multigraph and the goal is to delete the minimum number of arcs to ensure every strongly connected component of the resulting digraph is Eulerian. This problem is a natural extension of the Directed Feedback Arc Set problem and is also known to be motivated by certain scenarios arising in the study of housing markets. The complexity of the problem, when parameterized by solution size (i.e., size of the deletion set), has remained unresolved and has been highlighted in several papers. In this work, we answer this question by ruling out (subject to the usual complexity assumptions) a fixed-parameter algorithm (FPT algorithm) for this parameter and conduct a broad analysis of the problem with respect to other natural parameterizations. We prove both positive and negative results. Among these, we demonstrate that the problem is also hard (W[1]-hard or even para-NP-hard) when parameterized by either treewidth or maximum degree alone. Complementing our lower bounds, we establish that the problem is in XP when parameterized by treewidth and FPT when parameterized either by both treewidth and maximum degree or by both treewidth and solution size. We show that on simple digraphs, these algorithms have near-optimal asymptotic dependence on the treewidth assuming the Exponential Time Hypothesis.
{"title":"On the Parameterized Complexity of Eulerian Strong Component Arc Deletion","authors":"Václav Blažej, Satyabrata Jana, M. S. Ramanujan, Peter Strulo","doi":"10.1007/s00453-025-01336-6","DOIUrl":"10.1007/s00453-025-01336-6","url":null,"abstract":"<div><p>In this paper, we study the Eulerian Strong Component Arc Deletion problem, where the input is a directed multigraph and the goal is to delete the minimum number of arcs to ensure every strongly connected component of the resulting digraph is Eulerian. This problem is a natural extension of the Directed Feedback Arc Set problem and is also known to be motivated by certain scenarios arising in the study of housing markets. The complexity of the problem, when parameterized by solution size (i.e., size of the deletion set), has remained unresolved and has been highlighted in several papers. In this work, we answer this question by ruling out (subject to the usual complexity assumptions) a fixed-parameter algorithm (FPT algorithm) for this parameter and conduct a broad analysis of the problem with respect to other natural parameterizations. We prove both positive and negative results. Among these, we demonstrate that the problem is also hard (W[1]-hard or even para-NP-hard) when parameterized by either treewidth or maximum degree alone. Complementing our lower bounds, we establish that the problem is in XP when parameterized by treewidth and FPT when parameterized either by both treewidth and maximum degree or by both treewidth and solution size. We show that on simple digraphs, these algorithms have near-optimal asymptotic dependence on the treewidth assuming the Exponential Time Hypothesis.</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 11","pages":"1669 - 1709"},"PeriodicalIF":0.7,"publicationDate":"2025-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s00453-025-01336-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145090612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-02DOI: 10.1007/s00453-025-01321-z
Hans-Peter Lehmann, Peter Sanders, Stefan Walzer
A minimal perfect hash function (MPHF) maps a set S of n keys to the first n integers without collisions. There is a lower bound of (nlog _2e-mathcal {O}(log n) approx 1.44n) bits needed to represent an MPHF. This can be reached by a brute-force algorithm that tries (e^n) hash function seeds in expectation and stores the first seed that leads to an MPHF. The most space-efficient previous algorithms for constructing MPHFs all use such a brute-force approach as a basic building block. In this paper, we introduce ShockHash – Small, heavily overloaded cuckoo hash tables for minimal perfect hashing. ShockHash uses two hash functions (h_0) and (h_1), hoping for the existence of a function (f : S rightarrow {0,1}) such that (x mapsto h_{f(x)}(x)) is an MPHF on S. It then uses a 1-bit retrieval data structure to store f using (n + o(n)) bits. In graph terminology, ShockHash generates n-edge random graphs until stumbling on a pseudoforest – where each component contains as many edges as nodes. Using cuckoo hashing, ShockHash then derives an MPHF from the pseudoforest in linear time. We show that ShockHash needs to try only about ((e/2)^n approx 1.359^n) seeds in expectation. This reduces the space for storing the seed by roughly n bits (maintaining the asymptotically optimal space consumption) and speeds up construction by almost a factor of (2^n) compared to brute-force. Bipartite ShockHash reduces the expected construction time again to about (1.166^n) by maintaining a pool of candidate hash functions and checking all possible pairs. Using ShockHash as a building block within the RecSplit framework we obtain ShockHash-RS, which can be constructed up to 3 orders of magnitude faster than competing approaches. ShockHash-RS can build an MPHF for 10 million keys with 1.489 bits per key in about half an hour. When instead using ShockHash after an efficient k-perfect hash function, it achieves space usage similar to the best competitors, while being significantly faster to construct and query.
{"title":"ShockHash: Near Optimal-Space Minimal Perfect Hashing Beyond Brute-Force","authors":"Hans-Peter Lehmann, Peter Sanders, Stefan Walzer","doi":"10.1007/s00453-025-01321-z","DOIUrl":"10.1007/s00453-025-01321-z","url":null,"abstract":"<div><p>A minimal perfect hash function (MPHF) maps a set <i>S</i> of <i>n</i> keys to the first <i>n</i> integers without collisions. There is a lower bound of <span>(nlog _2e-mathcal {O}(log n) approx 1.44n)</span> bits needed to represent an MPHF. This can be reached by a <i>brute-force</i> algorithm that tries <span>(e^n)</span> hash function seeds in expectation and stores the first seed that leads to an MPHF. The most space-efficient previous algorithms for constructing MPHFs all use such a brute-force approach as a basic building block. In this paper, we introduce ShockHash – <b>S</b>mall, <b>h</b>eavily <b>o</b>verloaded cu<b>ck</b>oo <b>hash</b> tables for minimal perfect hashing. ShockHash uses two hash functions <span>(h_0)</span> and <span>(h_1)</span>, hoping for the existence of a function <span>(f : S rightarrow {0,1})</span> such that <span>(x mapsto h_{f(x)}(x))</span> is an MPHF on <i>S</i>. It then uses a 1-bit retrieval data structure to store <i>f</i> using <span>(n + o(n))</span> bits. In graph terminology, ShockHash generates <i>n</i>-edge random graphs until stumbling on a <i>pseudoforest</i> – where each component contains as many edges as nodes. Using cuckoo hashing, ShockHash then derives an MPHF from the pseudoforest in linear time. We show that ShockHash needs to try only about <span>((e/2)^n approx 1.359^n)</span> seeds in expectation. This reduces the space for storing the seed by roughly <i>n</i> bits (maintaining the asymptotically optimal space consumption) and speeds up construction by almost a factor of <span>(2^n)</span> compared to brute-force. <i>Bipartite</i> ShockHash reduces the expected construction time again to about <span>(1.166^n)</span> by maintaining a pool of candidate hash functions and checking all possible pairs. Using ShockHash as a building block within the RecSplit framework we obtain ShockHash-RS, which can be constructed up to 3 orders of magnitude faster than competing approaches. ShockHash-RS can build an MPHF for 10 million keys with 1.489 bits per key in about half an hour. When instead using ShockHash after an efficient <i>k</i>-perfect hash function, it achieves space usage similar to the best competitors, while being significantly faster to construct and query.</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 11","pages":"1620 - 1668"},"PeriodicalIF":0.7,"publicationDate":"2025-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s00453-025-01321-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145090312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-28DOI: 10.1007/s00453-025-01323-x
Andre Opris, Johannes Lengler, Dirk Sudholt
The (textsc {Jump} _k) benchmark was the first problem for which crossover was proven to give a speed-up over mutation-only evolutionary algorithms. Jansen and Wegener (Algorithmica 2002) proved an upper bound of (O(textrm{poly}(n) + 4^k/p_c)) for the ((mu )+1) Genetic Algorithm (((mu )+1) GA), but only for unrealistically small crossover probabilities (p_c). To this date, it remains an open problem to prove similar upper bounds for realistic (p_c); the best known runtime bound, in terms of function evaluations, for (p_c = Omega (1)) is (O((n/chi )^{k-1})), (chi ) a positive constant. We provide a novel approach and analyse the evolution of the population diversity, measured as sum of pairwise Hamming distances, for a variant of the ((mu )+1) GA on (textsc {Jump} _k). The ((mu )+1)-({lambda _c})-GA creates one offspring in each generation either by applying mutation to one parent or by applying crossover ({lambda _c}) times to the same two parents (followed by mutation), to amplify the probability of creating an accepted offspring in generations with crossover. We show that population diversity in the ((mu )+1)-({lambda _c})-GA converges to an equilibrium of near-perfect diversity. This yields an improved time bound of (O(mu n log (mu ) + 4^k)) function evaluations for a range of k under the mild assumptions (p_c = O(1/k)) and (mu in Omega (kn)). For all constant k, the restriction is satisfied for some (p_c = Omega (1)) and it implies that the expected runtime for all constant k and an appropriate (mu = Theta (kn)) is bounded by (O(n^2 log n)), irrespective of k. For larger k, the expected time of the ((mu )+1)-({lambda _c})-GA is (Theta (4^k)), which is tight for a large class of unbiased black-box algorithms and faster than the original ((mu )+1) GA by a factor of (Omega (1/p_c)). We also show that our analysis can be extended to other unitation functions such as (textsc {Jump} _{k, delta }) and Hurdle.
{"title":"Achieving Tight (O(4^k)) Runtime Bounds on Jumpk by Proving that Genetic Algorithms Evolve Near-Maximal Population Diversity","authors":"Andre Opris, Johannes Lengler, Dirk Sudholt","doi":"10.1007/s00453-025-01323-x","DOIUrl":"10.1007/s00453-025-01323-x","url":null,"abstract":"<div><p>The <span>(textsc {Jump} _k)</span> benchmark was the first problem for which crossover was proven to give a speed-up over mutation-only evolutionary algorithms. Jansen and Wegener (Algorithmica 2002) proved an upper bound of <span>(O(textrm{poly}(n) + 4^k/p_c))</span> for the (<span>(mu )</span>+1) Genetic Algorithm ((<span>(mu )</span>+1) GA), but only for unrealistically small crossover probabilities <span>(p_c)</span>. To this date, it remains an open problem to prove similar upper bounds for realistic <span>(p_c)</span>; the best known runtime bound, in terms of function evaluations, for <span>(p_c = Omega (1))</span> is <span>(O((n/chi )^{k-1}))</span>, <span>(chi )</span> a positive constant. We provide a novel approach and analyse the evolution of the population diversity, measured as sum of pairwise Hamming distances, for a variant of the (<span>(mu )</span>+1) GA on <span>(textsc {Jump} _k)</span>. The (<span>(mu )</span>+1)-<span>({lambda _c})</span>-GA creates one offspring in each generation either by applying mutation to one parent or by applying crossover <span>({lambda _c})</span> times to the same two parents (followed by mutation), to amplify the probability of creating an accepted offspring in generations with crossover. We show that population diversity in the (<span>(mu )</span>+1)-<span>({lambda _c})</span>-GA converges to an equilibrium of near-perfect diversity. This yields an improved time bound of <span>(O(mu n log (mu ) + 4^k))</span> function evaluations for a range of <i>k</i> under the mild assumptions <span>(p_c = O(1/k))</span> and <span>(mu in Omega (kn))</span>. For all constant <i>k</i>, the restriction is satisfied for some <span>(p_c = Omega (1))</span> and it implies that the expected runtime for all constant <i>k</i> and an appropriate <span>(mu = Theta (kn))</span> is bounded by <span>(O(n^2 log n))</span>, irrespective of <i>k</i>. For larger <i>k</i>, the expected time of the (<span>(mu )</span>+1)-<span>({lambda _c})</span>-GA is <span>(Theta (4^k))</span>, which is tight for a large class of unbiased black-box algorithms and faster than the original (<span>(mu )</span>+1) GA by a factor of <span>(Omega (1/p_c))</span>. We also show that our analysis can be extended to other unitation functions such as <span>(textsc {Jump} _{k, delta })</span> and H<span>urdle</span>.</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 11","pages":"1564 - 1619"},"PeriodicalIF":0.7,"publicationDate":"2025-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s00453-025-01323-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145090348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-21DOI: 10.1007/s00453-025-01335-7
Marvin Künnemann, Bodo Manthey, Rianne Veenstra
The 2-opt heuristic is a very simple local search heuristic for the traveling salesperson problem. In practice it usually converges quickly to solutions within a few percentages of optimality. In contrast to this, its running-time is exponential and its approximation performance is poor in the worst case. Englert, Röglin, and Vöcking (Algorithmica, 2014) provided a smoothed analysis in the so-called one-step model in order to explain the performance of 2-opt on d-dimensional Euclidean instances, both in terms of running-time and in terms of approximation ratio. However, translating their results to the classical model of smoothed analysis, where points are perturbed by Gaussian distributions with standard deviation (sigma ), yields only weak bounds. We prove bounds that are polynomial in n and (1/sigma ) for the smoothed running-time with Gaussian perturbations. In addition, our analysis for Euclidean distances is much simpler than the existing smoothed analysis. Furthermore, we prove a smoothed approximation ratio of (O(log (1/sigma ))). This bound is almost tight, as we also provide a lower bound of (Omega (frac{log n}{log log n})) for (sigma = O(1/sqrt{n})). Our main technical novelty here is that, different from existing smoothed analyses, we do not separately analyze objective values of the global and local optimum on all inputs (which only allows for a bound of (O(1/sigma ))), but simultaneously bound them on the same input.
{"title":"Smoothed Analysis of the 2-Opt Heuristic for the TSP under Gaussian Noise","authors":"Marvin Künnemann, Bodo Manthey, Rianne Veenstra","doi":"10.1007/s00453-025-01335-7","DOIUrl":"10.1007/s00453-025-01335-7","url":null,"abstract":"<div><p>The 2-opt heuristic is a very simple local search heuristic for the traveling salesperson problem. In practice it usually converges quickly to solutions within a few percentages of optimality. In contrast to this, its running-time is exponential and its approximation performance is poor in the worst case. Englert, Röglin, and Vöcking (<i>Algorithmica</i>, 2014) provided a smoothed analysis in the so-called one-step model in order to explain the performance of 2-opt on <i>d</i>-dimensional Euclidean instances, both in terms of running-time and in terms of approximation ratio. However, translating their results to the classical model of smoothed analysis, where points are perturbed by Gaussian distributions with standard deviation <span>(sigma )</span>, yields only weak bounds. We prove bounds that are polynomial in <i>n</i> and <span>(1/sigma )</span> for the smoothed running-time with Gaussian perturbations. In addition, our analysis for Euclidean distances is much simpler than the existing smoothed analysis. Furthermore, we prove a smoothed approximation ratio of <span>(O(log (1/sigma )))</span>. This bound is almost tight, as we also provide a lower bound of <span>(Omega (frac{log n}{log log n}))</span> for <span>(sigma = O(1/sqrt{n}))</span>. Our main technical novelty here is that, different from existing smoothed analyses, we do not separately analyze objective values of the global and local optimum on all inputs (which only allows for a bound of <span>(O(1/sigma ))</span>), but simultaneously bound them on the same input.</p></div>","PeriodicalId":50824,"journal":{"name":"Algorithmica","volume":"87 11","pages":"1518 - 1563"},"PeriodicalIF":0.7,"publicationDate":"2025-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s00453-025-01335-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145090695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}