Pub Date : 2025-02-28DOI: 10.1016/j.ipl.2025.106572
Jordan Dempsey , Leo van Iersel , Mark Jones , Norbert Zeh
Maximum agreement forests have been used as a measure of dissimilarity of two or more phylogenetic trees on a given set of taxa. An agreement forest is a set of trees that can be obtained from each of the input trees by deleting edges and suppressing degree-2 vertices. A maximum agreement forest is such a forest with the minimum number of components. We present a simple 4-approximation algorithm for computing a maximum agreement forest of multiple unrooted binary trees. This algorithm applies LP rounding to an extension of a recent ILP formulation of the maximum agreement forest problem on two trees by Van Wersch et al. [13]. We achieve the same approximation ratio as the algorithm by Chen et al. [3], but our algorithm is extremely simple. We also prove that no algorithm based on the ILP formulation by Van Wersch et al. can achieve an approximation ratio of , for any , even on two trees. To this end, we prove that the integrality gap of the ILP approaches 4 as the size of the two input trees grows.
{"title":"A simple 4-approximation algorithm for maximum agreement forests on multiple unrooted binary trees","authors":"Jordan Dempsey , Leo van Iersel , Mark Jones , Norbert Zeh","doi":"10.1016/j.ipl.2025.106572","DOIUrl":"10.1016/j.ipl.2025.106572","url":null,"abstract":"<div><div>Maximum agreement forests have been used as a measure of dissimilarity of two or more phylogenetic trees on a given set of taxa. An agreement forest is a set of trees that can be obtained from each of the input trees by deleting edges and suppressing degree-2 vertices. A maximum agreement forest is such a forest with the minimum number of components. We present a simple 4-approximation algorithm for computing a maximum agreement forest of multiple unrooted binary trees. This algorithm applies LP rounding to an extension of a recent ILP formulation of the maximum agreement forest problem on two trees by Van Wersch et al. <span><span>[13]</span></span>. We achieve the same approximation ratio as the algorithm by Chen et al. <span><span>[3]</span></span>, but our algorithm is extremely simple. We also prove that no algorithm based on the ILP formulation by Van Wersch et al. can achieve an approximation ratio of <span><math><mn>4</mn><mo>−</mo><mi>ε</mi></math></span>, for any <span><math><mi>ε</mi><mo>></mo><mn>0</mn></math></span>, even on two trees. To this end, we prove that the integrality gap of the ILP approaches 4 as the size of the two input trees grows.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"190 ","pages":"Article 106572"},"PeriodicalIF":0.7,"publicationDate":"2025-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143534765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-02-15DOI: 10.1016/j.ipl.2025.106571
Jan Kleinekathöfer, Alireza Mahzoon, Rolf Drechsler
Decision Diagrams (DDs) are among the most popular representations for Boolean functions. They are widely used in the synthesis and verification of digital circuits. The size (i.e., number of nodes) and computation time (required time for performing operations) are two important parameters that determine the efficiency of a DD in different applications. It has been proven that some DDs can represent specific functions in polynomial space or perform certain operations in polynomial time. For example, Binary Decision Diagrams (BDDs) are capable of representing a wide variety of functions (e.g. integer addition) in polynomial space with respect to the input size. However, there are also some functions (e.g., integer multiplication) for which the exponential lower-bounds have been proven for the BDD sizes.
In this paper, we investigate the space complexity of representing an integer addition, where one of the operands is shifted to the right by an arbitrary value. We call this function the shifted addition. This function is widely used in many digital circuits, e.g., floating point adders. We prove that the size of the BDD representing a shifted addition has exponential space complexity with respect to the input size. It is an important step towards clarifying the reasons behind the failure of BDD-based verification and synthesis when they are applied to the circuits containing shifted addition, e.g., floating point adders.
{"title":"Lower bound proof for the size of BDDs representing a shifted addition","authors":"Jan Kleinekathöfer, Alireza Mahzoon, Rolf Drechsler","doi":"10.1016/j.ipl.2025.106571","DOIUrl":"10.1016/j.ipl.2025.106571","url":null,"abstract":"<div><div><em>Decision Diagrams</em> (DDs) are among the most popular representations for Boolean functions. They are widely used in the synthesis and verification of digital circuits. The size (i.e., number of nodes) and computation time (required time for performing operations) are two important parameters that determine the efficiency of a DD in different applications. It has been proven that some DDs can represent specific functions in polynomial space or perform certain operations in polynomial time. For example, <em>Binary Decision Diagrams</em> (BDDs) are capable of representing a wide variety of functions (e.g. integer addition) in polynomial space with respect to the input size. However, there are also some functions (e.g., integer multiplication) for which the exponential lower-bounds have been proven for the BDD sizes.</div><div>In this paper, we investigate the space complexity of representing an integer addition, where one of the operands is shifted to the right by an arbitrary value. We call this function the shifted addition. This function is widely used in many digital circuits, e.g., floating point adders. We prove that the size of the BDD representing a shifted addition has exponential space complexity with respect to the input size. It is an important step towards clarifying the reasons behind the failure of BDD-based verification and synthesis when they are applied to the circuits containing shifted addition, e.g., floating point adders.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"190 ","pages":"Article 106571"},"PeriodicalIF":0.7,"publicationDate":"2025-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143427877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-02-14DOI: 10.1016/j.ipl.2025.106570
Dekel Tsur
In the Cliques or Trees Vertex Deletion problem, the input is a graph G and an integer k, and the goal is to decide whether there is a set of at most k vertices whose removal from G result in a graph in which every connected component is either a clique or a tree. In this paper we give an -time deterministic algorithm, an -time randomized algorithm, and a kernel with vertices for Cliques or Trees Vertex Deletion.
{"title":"Faster algorithms and a smaller kernel for Cliques or Trees Vertex Deletion","authors":"Dekel Tsur","doi":"10.1016/j.ipl.2025.106570","DOIUrl":"10.1016/j.ipl.2025.106570","url":null,"abstract":"<div><div>In the <span>Cliques or Trees Vertex Deletion</span> problem, the input is a graph <em>G</em> and an integer <em>k</em>, and the goal is to decide whether there is a set of at most <em>k</em> vertices whose removal from <em>G</em> result in a graph in which every connected component is either a clique or a tree. In this paper we give an <span><math><msup><mrow><mi>O</mi></mrow><mrow><mo>⁎</mo></mrow></msup><mo>(</mo><msup><mrow><mn>3.46</mn></mrow><mrow><mi>k</mi></mrow></msup><mo>)</mo></math></span>-time deterministic algorithm, an <span><math><msup><mrow><mi>O</mi></mrow><mrow><mo>⁎</mo></mrow></msup><mo>(</mo><msup><mrow><mn>3.103</mn></mrow><mrow><mi>k</mi></mrow></msup><mo>)</mo></math></span>-time randomized algorithm, and a kernel with <span><math><mi>O</mi><mo>(</mo><msup><mrow><mi>k</mi></mrow><mrow><mn>4</mn></mrow></msup><mo>)</mo></math></span> vertices for <span>Cliques or Trees Vertex Deletion</span>.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"190 ","pages":"Article 106570"},"PeriodicalIF":0.7,"publicationDate":"2025-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143420534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-02-11DOI: 10.1016/j.ipl.2025.106569
Daniel Eichhorn, Sven O. Krumke
The Online Delay Management on a Single Train Line (ODMP) deals with the question at which station a train should wait for delayed passengers, instead of forcing them to take the next train. Waiting at a station increases the delay of all passengers that are already on board, and the goal is to minimize the total passenger delay. An online algorithm learns about the number of delayed passengers at a station only when reaching this station. We study the ODMP with an additional prediction on the future input data, which an online algorithm can utilize. Two desired qualities for online algorithms with prediction are called consistency and robustness, denoting the competitive ratio in case of best and worst prediction respectively. We present a family of algorithms, which uses a hyperparameter measuring the “doubt” about the given prediction. This allows to achieve -consistency and -robustness. Moreover, we provide a lower bound for the trade-off between consistency and robustness for two variously detailed prediction models, showing that our algorithm achieves an asymptotically optimal trade-off for small values of λ.
{"title":"Online delay management on a single train line with predictions","authors":"Daniel Eichhorn, Sven O. Krumke","doi":"10.1016/j.ipl.2025.106569","DOIUrl":"10.1016/j.ipl.2025.106569","url":null,"abstract":"<div><div>The Online Delay Management on a Single Train Line (ODMP) deals with the question at which station a train should wait for delayed passengers, instead of forcing them to take the next train. Waiting at a station increases the delay of all passengers that are already on board, and the goal is to minimize the total passenger delay. An online algorithm learns about the number of delayed passengers at a station only when reaching this station. We study the ODMP with an additional prediction on the future input data, which an online algorithm can utilize. Two desired qualities for online algorithms with prediction are called consistency and robustness, denoting the competitive ratio in case of best and worst prediction respectively. We present a family of algorithms, which uses a hyperparameter <span><math><mi>λ</mi><mo>∈</mo><mo>(</mo><mn>0</mn><mo>,</mo><mn>1</mn><mo>)</mo></math></span> measuring the “doubt” about the given prediction. This allows to achieve <span><math><mo>(</mo><mn>1</mn><mo>+</mo><mi>λ</mi><mo>)</mo></math></span>-consistency and <span><math><mo>(</mo><mn>1</mn><mo>+</mo><mn>1</mn><mo>/</mo><mi>λ</mi><mo>)</mo></math></span>-robustness. Moreover, we provide a lower bound for the trade-off between consistency and robustness for two variously detailed prediction models, showing that our algorithm achieves an asymptotically optimal trade-off for small values of <em>λ</em>.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"190 ","pages":"Article 106569"},"PeriodicalIF":0.7,"publicationDate":"2025-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143420533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-21DOI: 10.1016/j.ipl.2025.106560
Arnab Bhattacharyya , Sutanu Gayen , Kuldeep S. Meel , Dimitrios Myrisiotis , A. Pavan , N.V. Vinodchandran
We show that computing the total variation distance between two product distributions is -complete. This is in stark contrast with other distance measures such as Kullback–Leibler, Chi-square, and Hellinger, which tensorize over the marginals leading to efficient algorithms.
我们证明计算两个产品分布之间的总变异距离是# p -完备的。这与其他距离度量(如Kullback-Leibler、Chi-square和Hellinger)形成鲜明对比,后者在边缘上张紧,从而产生高效的算法。
{"title":"Total variation distance for product distributions is #P-complete","authors":"Arnab Bhattacharyya , Sutanu Gayen , Kuldeep S. Meel , Dimitrios Myrisiotis , A. Pavan , N.V. Vinodchandran","doi":"10.1016/j.ipl.2025.106560","DOIUrl":"10.1016/j.ipl.2025.106560","url":null,"abstract":"<div><div>We show that computing the total variation distance between two product distributions is <span><math><mi>#</mi><mi>P</mi></math></span>-complete. This is in stark contrast with other distance measures such as Kullback–Leibler, Chi-square, and Hellinger, which tensorize over the marginals leading to efficient algorithms.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"189 ","pages":"Article 106560"},"PeriodicalIF":0.7,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143099054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This work examines the Conditional Approval Framework for elections involving multiple interdependent issues, specifically focusing on the Conditional Minisum Approval Voting Rule. We first conduct a detailed analysis of the computational complexity of this rule, demonstrating that no approach can significantly outperform the brute-force algorithm under common computational complexity assumptions and various natural input restrictions. In response, we propose two practical restrictions (the first in the literature) that make the problem computationally tractable and show that these restrictions are essentially tight. Overall, this work provides a clear picture of the tractability landscape of the problem, contributing to a comprehensive understanding of the complications introduced by conditional ballots and indicating that conditional approval voting can be applied in practice, albeit under specific conditions.
{"title":"On the Tractability Landscape of the Conditional Minisum Approval Voting Rule","authors":"Georgios Amanatidis , Michael Lampis , Evangelos Markakis , Georgios Papasotiropoulos","doi":"10.1016/j.ipl.2025.106561","DOIUrl":"10.1016/j.ipl.2025.106561","url":null,"abstract":"<div><div>This work examines the Conditional Approval Framework for elections involving multiple interdependent issues, specifically focusing on the Conditional Minisum Approval Voting Rule. We first conduct a detailed analysis of the computational complexity of this rule, demonstrating that no approach can significantly outperform the brute-force algorithm under common computational complexity assumptions and various natural input restrictions. In response, we propose two practical restrictions (the first in the literature) that make the problem computationally tractable and show that these restrictions are essentially tight. Overall, this work provides a clear picture of the tractability landscape of the problem, contributing to a comprehensive understanding of the complications introduced by conditional ballots and indicating that conditional approval voting can be applied in practice, albeit under specific conditions.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"189 ","pages":"Article 106561"},"PeriodicalIF":0.7,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143099053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-16DOI: 10.1016/j.ipl.2025.106559
Alexandros A. Voudouris
We consider a distributed voting problem with a set of agents that are partitioned into disjoint groups and a set of obnoxious alternatives. Agents and alternatives are represented by points in a metric space. The goal is to compute the alternative that maximizes the total distance from all agents using a two-step mechanism which, given some information about the distances between agents and alternatives, first chooses a representative alternative for each group of agents, and then declares one of them as the overall winner. Due to the restricted nature of the mechanism and the potentially limited information it has to make its decision, it might not be always possible to choose the optimal alternative. We show tight bounds on the distortion of different mechanisms depending on the amount of the information they have access to; in particular, we study full-information and ordinal mechanisms.
{"title":"Metric distortion of obnoxious distributed voting","authors":"Alexandros A. Voudouris","doi":"10.1016/j.ipl.2025.106559","DOIUrl":"10.1016/j.ipl.2025.106559","url":null,"abstract":"<div><div>We consider a distributed voting problem with a set of agents that are partitioned into disjoint groups and a set of <em>obnoxious</em> alternatives. Agents and alternatives are represented by points in a metric space. The goal is to compute the alternative that <em>maximizes</em> the total distance from all agents using a two-step mechanism which, given some information about the distances between agents and alternatives, first chooses a representative alternative for each group of agents, and then declares one of them as the overall winner. Due to the restricted nature of the mechanism and the potentially limited information it has to make its decision, it might not be always possible to choose the optimal alternative. We show tight bounds on the <em>distortion</em> of different mechanisms depending on the amount of the information they have access to; in particular, we study full-information and ordinal mechanisms.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"189 ","pages":"Article 106559"},"PeriodicalIF":0.7,"publicationDate":"2025-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143156875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-14DOI: 10.1016/j.ipl.2025.106558
Michael T. Goodrich
The Quickhull algorithm is a simple algorithm for constructing the convex hull of a set of n points. Quickhull is usually described for points in the plane, in which case it is defined as a divide-and-conquer algorithm, where one has a pair of points such that p and r are on the convex hull, and one then finds the point, q, farthest from the line , which must also be on the convex hull, and then uses the triangle to divide the remaining points and recursively solve the resulting subproblems. It is well-known that Quickhull has a worst-case running time of , but it runs much faster than this for some input distributions. In a highly cited paper, Barber, Dobkin, and Huhdanpaa conjecture that the Quickhull algorithm runs in worst-case time, where h is the size of the convex hull, when the input points have precision . In this paper, we give an explicit lower-bound construction that shows that, in general, the worst-case running time of the Quickhull algorithm is . Our lower bound proof also provides a counter-example to the Quickhull precision conjecture of Barber et al., in that we give an explicit construction of a set, S, of n points with precision such that h is but the worst-case running time of Quickhull on S is , not .
{"title":"A lower bound for the Quickhull convex hull algorithm that disproves the Quickhull precision conjecture","authors":"Michael T. Goodrich","doi":"10.1016/j.ipl.2025.106558","DOIUrl":"10.1016/j.ipl.2025.106558","url":null,"abstract":"<div><div>The <em><strong>Quickhull</strong></em> algorithm is a simple algorithm for constructing the convex hull of a set of <em>n</em> points. Quickhull is usually described for points in the plane, in which case it is defined as a divide-and-conquer algorithm, where one has a pair of points <span><math><mo>(</mo><mi>p</mi><mo>,</mo><mi>r</mi><mo>)</mo></math></span> such that <em>p</em> and <em>r</em> are on the convex hull, and one then finds the point, <em>q</em>, farthest from the line <span><math><mover><mrow><mi>p</mi><mi>r</mi></mrow><mo>‾</mo></mover></math></span>, which must also be on the convex hull, and then uses the triangle <span><math><mo>(</mo><mi>p</mi><mo>,</mo><mi>q</mi><mo>,</mo><mi>r</mi><mo>)</mo></math></span> to divide the remaining points and recursively solve the resulting subproblems. It is well-known that Quickhull has a worst-case running time of <span><math><mi>Θ</mi><mo>(</mo><msup><mrow><mi>n</mi></mrow><mrow><mn>2</mn></mrow></msup><mo>)</mo></math></span>, but it runs much faster than this for some input distributions. In a highly cited paper, Barber, Dobkin, and Huhdanpaa conjecture that the Quickhull algorithm runs in worst-case <span><math><mi>O</mi><mo>(</mo><mi>n</mi><mi>log</mi><mo></mo><mi>h</mi><mo>)</mo></math></span> time, where <em>h</em> is the size of the convex hull, when the input points have precision <span><math><mi>O</mi><mo>(</mo><mi>log</mi><mo></mo><mi>n</mi><mo>)</mo></math></span>. In this paper, we give an explicit lower-bound construction that shows that, in general, the worst-case running time of the Quickhull algorithm is <span><math><mi>Θ</mi><mo>(</mo><mi>n</mi><mi>h</mi><mo>)</mo></math></span>. Our lower bound proof also provides a counter-example to the Quickhull precision conjecture of Barber et al., in that we give an explicit construction of a set, <em>S</em>, of <em>n</em> points with precision <span><math><mi>O</mi><mo>(</mo><mi>log</mi><mo></mo><mi>n</mi><mo>)</mo></math></span> such that <em>h</em> is <span><math><mi>O</mi><mo>(</mo><mi>log</mi><mo></mo><mi>n</mi><mo>)</mo></math></span> but the worst-case running time of Quickhull on <em>S</em> is <span><math><mi>Θ</mi><mo>(</mo><mi>n</mi><mi>h</mi><mo>)</mo></math></span>, not <span><math><mi>O</mi><mo>(</mo><mi>n</mi><mi>log</mi><mo></mo><mi>h</mi><mo>)</mo></math></span>.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"189 ","pages":"Article 106558"},"PeriodicalIF":0.7,"publicationDate":"2025-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143099051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-13DOI: 10.1016/j.ipl.2025.106557
Tamanna Chhabra , Sukhpal Singh Ghuman , Jorma Tarhio
We present new algorithms for the k mismatches version of approximate string matching. Our algorithms utilize the SIMD (Single Instruction Multiple Data) instruction set extensions, particularly AVX2 and AVX-512 instructions. Our approach is an extension of an earlier algorithm for exact string matching with SSE2 and AVX2. In addition, we modify this exact string matching algorithm to work with AVX-512. We demonstrate the competitiveness of our solutions by practical experiments. Our algorithms outperform earlier algorithms for both exact and approximate string matching on various benchmark data sets.
{"title":"String searching with mismatches using AVX2 and AVX-512 instructions","authors":"Tamanna Chhabra , Sukhpal Singh Ghuman , Jorma Tarhio","doi":"10.1016/j.ipl.2025.106557","DOIUrl":"10.1016/j.ipl.2025.106557","url":null,"abstract":"<div><div>We present new algorithms for the <em>k</em> mismatches version of approximate string matching. Our algorithms utilize the SIMD (Single Instruction Multiple Data) instruction set extensions, particularly AVX2 and AVX-512 instructions. Our approach is an extension of an earlier algorithm for exact string matching with SSE2 and AVX2. In addition, we modify this exact string matching algorithm to work with AVX-512. We demonstrate the competitiveness of our solutions by practical experiments. Our algorithms outperform earlier algorithms for both exact and approximate string matching on various benchmark data sets.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"189 ","pages":"Article 106557"},"PeriodicalIF":0.7,"publicationDate":"2025-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143099052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-03DOI: 10.1016/j.ipl.2024.106556
Naoto Ohsaka
Given a two-prover game G and its two satisfying labelings and , the Label Cover Reconfiguration problem asks whether can be transformed into by repeatedly changing the label of a single vertex while preserving any intermediate labeling satisfying G. We consider its optimization version by relaxing the feasibility of labelings, referred to as Maxmin Label Cover Reconfiguration: We are allowed to pass through any non-satisfying labelings, but required to maximize the “soundness error,” which is defined as the minimum fraction of satisfied edges during transformation from to . Since the parallel repetition theorem of Raz (1998) [32], which implies -hardness of approximating Label Cover within any constant factor, gives strong inapproximability results for many -hard problems, one may think of using Maxmin Label Cover Reconfiguration to derive inapproximability results for reconfiguration problems. We prove the following results on Maxmin Label Cover Reconfiguration, which display different trends from those of Label Cover and the parallel repetition theorem:
•
Maxmin Label Cover Reconfiguration can be approximated within a factor of for some restricted graph classes, including biregular graphs, balanced bipartite graphs with no isolated vertices, and superconstant average degree graphs.
•
A “naive” parallel repetition of Maxmin Label Cover Reconfiguration does not decrease the soundness error for every two-prover game.
•
Label Cover Reconfiguration on projection games can be decided in polynomial time.
Our results suggest that a reconfiguration analogue of the parallel repetition theorem is unlikely.
{"title":"On approximate reconfigurability of label cover","authors":"Naoto Ohsaka","doi":"10.1016/j.ipl.2024.106556","DOIUrl":"10.1016/j.ipl.2024.106556","url":null,"abstract":"<div><div>Given a two-prover game <em>G</em> and its two satisfying labelings <span><math><msub><mrow><mi>ψ</mi></mrow><mrow><mi>ini</mi></mrow></msub></math></span> and <span><math><msub><mrow><mi>ψ</mi></mrow><mrow><mi>tar</mi></mrow></msub></math></span>, the <span>Label Cover Reconfiguration</span> problem asks whether <span><math><msub><mrow><mi>ψ</mi></mrow><mrow><mi>ini</mi></mrow></msub></math></span> can be transformed into <span><math><msub><mrow><mi>ψ</mi></mrow><mrow><mi>tar</mi></mrow></msub></math></span> by repeatedly changing the label of a single vertex while preserving any intermediate labeling satisfying <em>G</em>. We consider its optimization version by relaxing the feasibility of labelings, referred to as <span>Maxmin Label Cover Reconfiguration</span>: We are allowed to pass through any <em>non-satisfying</em> labelings, but required to maximize the “soundness error,” which is defined as the <em>minimum</em> fraction of satisfied edges during transformation from <span><math><msub><mrow><mi>ψ</mi></mrow><mrow><mi>ini</mi></mrow></msub></math></span> to <span><math><msub><mrow><mi>ψ</mi></mrow><mrow><mi>tar</mi></mrow></msub></math></span>. Since the parallel repetition theorem of Raz (1998) <span><span>[32]</span></span>, which implies <figure><img></figure>-hardness of approximating <span>Label Cover</span> within any constant factor, gives strong inapproximability results for many <figure><img></figure>-hard problems, one may think of using <span>Maxmin Label Cover Reconfiguration</span> to derive inapproximability results for reconfiguration problems. We prove the following results on <span>Maxmin Label Cover Reconfiguration</span>, which display different trends from those of <span>Label Cover</span> and the parallel repetition theorem:<ul><li><span>•</span><span><div><span>Maxmin Label Cover Reconfiguration</span> can be approximated within a factor of <span><math><mfrac><mrow><mn>1</mn></mrow><mrow><mn>4</mn></mrow></mfrac><mo>−</mo><mi>o</mi><mo>(</mo><mn>1</mn><mo>)</mo></math></span> for some restricted graph classes, including biregular graphs, balanced bipartite graphs with no isolated vertices, and superconstant average degree graphs.</div></span></li><li><span>•</span><span><div>A “naive” parallel repetition of <span>Maxmin Label Cover Reconfiguration</span> does not decrease the soundness error for <em>every</em> two-prover game.</div></span></li><li><span>•</span><span><div><span>Label Cover Reconfiguration</span> on <em>projection games</em> can be decided in polynomial time.</div></span></li></ul> Our results suggest that a reconfiguration analogue of the parallel repetition theorem is unlikely.</div></div>","PeriodicalId":56290,"journal":{"name":"Information Processing Letters","volume":"189 ","pages":"Article 106556"},"PeriodicalIF":0.7,"publicationDate":"2025-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143099050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}