首页 > 最新文献

Bulletin of the Society of Sea Water Science, Japan最新文献

英文 中文
Computing Maximal Unique Matches with the r-index 用r索引计算最大唯一匹配
Pub Date : 2022-05-03 DOI: 10.48550/arXiv.2205.01576
Sara Giuliani, Giuseppe Romana, Massimiliano Rossi
In recent years, pangenomes received increasing attention from the scientific community for their ability to incorporate population variation information and alleviate reference genome bias. Maximal Exact Matches ( MEMs ) and Maximal Unique Matches ( MUMs ) have proven themselves to be useful in multiple bioinformatic contexts, for example short-read alignment and multiple-genome alignment. However, standard techniques using suffix trees and FM-indexes do not scale to a pangenomic level. Recently, Gagie et al. [JACM 20] introduced the r -index that is a Burrows-Wheeler Transform ( BWT )-based index able to handle hundreds of human genomes. Later, Rossi et al. [JCB 22] enabled the computation of MEMs using the r -index, and Boucher et al. [DCC 21] showed how to compute them in a streaming fashion. In this paper, we show how to augment Boucher et al.’s approach to enable the computation of MUMs on the r -index, while preserving the space and time bounds. We add additional O ( r ) samples of the longest common prefix ( LCP ) array, where r is the number of equal-letter runs of the BWT , that permits the computation of the second longest match of the pattern suffix with respect to the input text, which in turn allows the computation of candidate MUMs . We implemented a proof-of-concept of our approach, that we call mum-phinder , and tested on real-world datasets. We compared our approach with competing methods that are able to compute MUMs . We observe that our method is up to 8 times smaller, while up to 19 times slower when the dataset is not highly repetitive, while on highly repetitive data, our method is up to 6.5 times slower and uses up to 25 times less memory.
近年来,泛基因组因其整合种群变异信息和减轻参考基因组偏差的能力而受到科学界的越来越多的关注。最大精确匹配(MEMs)和最大唯一匹配(mum)已被证明在多种生物信息学背景下非常有用,例如短读比对和多基因组比对。然而,使用后缀树和fm索引的标准技术不能扩展到全基因组水平。最近,Gagie等人[JACM 20]引入了r -索引,这是一种基于Burrows-Wheeler变换(BWT)的索引,能够处理数百个人类基因组。后来,Rossi等人[JCB 22]使用r -指数实现了MEMs的计算,而Boucher等人[DCC 21]展示了如何以流方式计算它们。在本文中,我们展示了如何增强Boucher等人的方法来实现r -索引上的mum计算,同时保留空间和时间界限。我们添加了最长公共前缀(LCP)数组的额外O (r)个样本,其中r是BWT的等字母运行次数,它允许计算模式后缀相对于输入文本的第二长的匹配,这反过来允许计算候选的MUMs。我们实现了我们的方法的概念验证,我们称之为mum-phinder,并在现实世界的数据集上进行了测试。我们将我们的方法与能够计算mom的竞争方法进行了比较。我们观察到,当数据集不是高度重复的时候,我们的方法要小8倍,而慢19倍,而在高度重复的数据上,我们的方法要慢6.5倍,使用的内存要少25倍。
{"title":"Computing Maximal Unique Matches with the r-index","authors":"Sara Giuliani, Giuseppe Romana, Massimiliano Rossi","doi":"10.48550/arXiv.2205.01576","DOIUrl":"https://doi.org/10.48550/arXiv.2205.01576","url":null,"abstract":"In recent years, pangenomes received increasing attention from the scientific community for their ability to incorporate population variation information and alleviate reference genome bias. Maximal Exact Matches ( MEMs ) and Maximal Unique Matches ( MUMs ) have proven themselves to be useful in multiple bioinformatic contexts, for example short-read alignment and multiple-genome alignment. However, standard techniques using suffix trees and FM-indexes do not scale to a pangenomic level. Recently, Gagie et al. [JACM 20] introduced the r -index that is a Burrows-Wheeler Transform ( BWT )-based index able to handle hundreds of human genomes. Later, Rossi et al. [JCB 22] enabled the computation of MEMs using the r -index, and Boucher et al. [DCC 21] showed how to compute them in a streaming fashion. In this paper, we show how to augment Boucher et al.’s approach to enable the computation of MUMs on the r -index, while preserving the space and time bounds. We add additional O ( r ) samples of the longest common prefix ( LCP ) array, where r is the number of equal-letter runs of the BWT , that permits the computation of the second longest match of the pattern suffix with respect to the input text, which in turn allows the computation of candidate MUMs . We implemented a proof-of-concept of our approach, that we call mum-phinder , and tested on real-world datasets. We compared our approach with competing methods that are able to compute MUMs . We observe that our method is up to 8 times smaller, while up to 19 times slower when the dataset is not highly repetitive, while on highly repetitive data, our method is up to 6.5 times slower and uses up to 25 times less memory.","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"35 1","pages":"22:1-22:16"},"PeriodicalIF":0.0,"publicationDate":"2022-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73530300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Parallel Flow-Based Hypergraph Partitioning 基于并行流的超图分区
Pub Date : 2022-01-05 DOI: 10.4230/LIPIcs.SEA.2022.5
Lars Gottesbüren, Tobias Heuer, P. Sanders
We present a shared-memory parallelization of flow-based refinement , which is considered the most powerful iterative improvement technique for hypergraph partitioning at the moment. Flow-based refinement works on bipartitions, so current sequential partitioners schedule it on different block pairs to improve k -way partitions. We investigate two different sources of parallelism: a parallel scheduling scheme and a parallel maximum flow algorithm based on the well-known push-relabel algorithm. In addition to thoroughly engineered implementations, we propose several optimizations that substantially accelerate the algorithm in practice, enabling the use on extremely large hypergraphs (up to 1 billion pins). We integrate our approach in the state-of-the-art parallel multilevel framework Mt-KaHyPar and conduct extensive experiments on a benchmark set of more than 500 real-world hypergraphs, to show that the partition quality of our code is on par with the highest quality sequential code ( KaHyPar ), while being an order of magnitude faster with 10 threads. .
我们提出了一种基于流的共享内存并行化优化方法,它被认为是目前超图划分最强大的迭代改进技术。基于流的细化工作在双分区上,所以当前的顺序分区者在不同的块对上调度它来改进k -way分区。我们研究了两种不同的并行性来源:一种并行调度方案和一种基于著名的推标签算法的并行最大流量算法。除了彻底的工程实现之外,我们还提出了几个优化,这些优化在实践中大大加快了算法的速度,使其能够在超大的超图(高达10亿个引脚)上使用。我们将我们的方法集成到最先进的并行多层框架Mt-KaHyPar中,并在超过500个真实超图的基准集上进行了广泛的实验,以表明我们代码的分区质量与最高质量的顺序代码(KaHyPar)相当,同时在10个线程时速度要快一个数量级。
{"title":"Parallel Flow-Based Hypergraph Partitioning","authors":"Lars Gottesbüren, Tobias Heuer, P. Sanders","doi":"10.4230/LIPIcs.SEA.2022.5","DOIUrl":"https://doi.org/10.4230/LIPIcs.SEA.2022.5","url":null,"abstract":"We present a shared-memory parallelization of flow-based refinement , which is considered the most powerful iterative improvement technique for hypergraph partitioning at the moment. Flow-based refinement works on bipartitions, so current sequential partitioners schedule it on different block pairs to improve k -way partitions. We investigate two different sources of parallelism: a parallel scheduling scheme and a parallel maximum flow algorithm based on the well-known push-relabel algorithm. In addition to thoroughly engineered implementations, we propose several optimizations that substantially accelerate the algorithm in practice, enabling the use on extremely large hypergraphs (up to 1 billion pins). We integrate our approach in the state-of-the-art parallel multilevel framework Mt-KaHyPar and conduct extensive experiments on a benchmark set of more than 500 real-world hypergraphs, to show that the partition quality of our code is on par with the highest quality sequential code ( KaHyPar ), while being an order of magnitude faster with 10 threads. .","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"48 1","pages":"5:1-5:21"},"PeriodicalIF":0.0,"publicationDate":"2022-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85796554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
An Adaptive Refinement Algorithm for Discretizations of Nonconvex QCQP 非凸QCQP离散化的自适应细化算法
Pub Date : 2022-01-01 DOI: 10.4230/LIPIcs.SEA.2022.24
A. Gupte, A. Koster, Sascha Kuhnke
We present an iterative algorithm to compute feasible solutions in reasonable running time to quadratically constrained quadratic programs (QCQPs), which form a challenging class of nonconvex continuous optimization. This algorithm is based on a mixed-integer linear program (MILP) which is a restriction of the original QCQP obtained by discretizing all quadratic terms. In each iteration, this MILP restriction is solved to get a feasible QCQP solution. Since the quality of this solution heavily depends on the chosen discretization of the MILP, we iteratively adapt the discretization values based on the MILP solution of the previous iteration. To maintain a reasonable problem size in each iteration of the algorithm, the discretization sizes are fixed at predefined values. Although our algorithm did not always yield good feasible solutions on arbitrary QCQP instances, an extensive computational study on almost 1300 test instances of two different problem classes – box-constrained quadratic programs with complementarity constraints and disjoint bilinear programs, demonstrates the effectiveness of our approach. We compare the quality of our solutions against those from heuristics and local optimization algorithms in two state-of-the-art commercial solvers and observe that on one instance class we clearly outperform the other methods whereas on the other class we obtain competitive results.
二次约束二次规划(QCQPs)是一类具有挑战性的非凸连续优化问题,本文提出了一种在合理运行时间内计算可行解的迭代算法。该算法基于混合整数线性规划(MILP), MILP是对原QCQP的约束,原QCQP是由所有二次项离散得到的。在每次迭代中,求解该MILP约束,得到可行的QCQP解。由于该解的质量很大程度上取决于所选择的离散化的MILP,我们迭代地适应离散化值基于前迭代的MILP解。为了在算法的每次迭代中保持合理的问题大小,将离散化大小固定在预定义的值。尽管我们的算法并不总是在任意QCQP实例上产生良好的可行解,但对两种不同问题类(具有互补约束的盒约束二次规划和不相交双线性规划)的近1300个测试实例进行了广泛的计算研究,证明了我们的方法的有效性。我们将我们的解决方案的质量与两个最先进的商业解决方案中的启发式和局部优化算法的解决方案进行了比较,并观察到在一个实例类上我们明显优于其他方法,而在另一个实例类上我们获得了竞争结果。
{"title":"An Adaptive Refinement Algorithm for Discretizations of Nonconvex QCQP","authors":"A. Gupte, A. Koster, Sascha Kuhnke","doi":"10.4230/LIPIcs.SEA.2022.24","DOIUrl":"https://doi.org/10.4230/LIPIcs.SEA.2022.24","url":null,"abstract":"We present an iterative algorithm to compute feasible solutions in reasonable running time to quadratically constrained quadratic programs (QCQPs), which form a challenging class of nonconvex continuous optimization. This algorithm is based on a mixed-integer linear program (MILP) which is a restriction of the original QCQP obtained by discretizing all quadratic terms. In each iteration, this MILP restriction is solved to get a feasible QCQP solution. Since the quality of this solution heavily depends on the chosen discretization of the MILP, we iteratively adapt the discretization values based on the MILP solution of the previous iteration. To maintain a reasonable problem size in each iteration of the algorithm, the discretization sizes are fixed at predefined values. Although our algorithm did not always yield good feasible solutions on arbitrary QCQP instances, an extensive computational study on almost 1300 test instances of two different problem classes – box-constrained quadratic programs with complementarity constraints and disjoint bilinear programs, demonstrates the effectiveness of our approach. We compare the quality of our solutions against those from heuristics and local optimization algorithms in two state-of-the-art commercial solvers and observe that on one instance class we clearly outperform the other methods whereas on the other class we obtain competitive results.","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"30 1","pages":"24:1-24:14"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74991349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Practical Performance of Random Projections in Linear Programming 线性规划中随机投影的实际性能
Pub Date : 2022-01-01 DOI: 10.4230/LIPIcs.SEA.2022.21
Leo Liberti, Benedetto Manca, Pierre-Louis Poirion
The use of random projections in mathematical programming allows standard solution algorithms to solve instances of much larger sizes, at least approximately. Approximation results have been derived in the relevant literature for many specific problems, as well as for several mathematical programming subclasses. Despite the theoretical developments, it is not always clear that random projections are actually useful in solving mathematical programs in practice. In this paper we provide a computational assessment of the application of random projections to linear programming.
在数学规划中使用随机投影允许标准求解算法求解更大的实例,至少是近似的。对于许多具体问题,以及一些数学规划子类,相关文献已经导出了近似结果。尽管在理论上有了发展,但在实践中,随机投影在解决数学程序中是否真的有用并不总是很清楚。本文给出了随机投影在线性规划中的应用。
{"title":"Practical Performance of Random Projections in Linear Programming","authors":"Leo Liberti, Benedetto Manca, Pierre-Louis Poirion","doi":"10.4230/LIPIcs.SEA.2022.21","DOIUrl":"https://doi.org/10.4230/LIPIcs.SEA.2022.21","url":null,"abstract":"The use of random projections in mathematical programming allows standard solution algorithms to solve instances of much larger sizes, at least approximately. Approximation results have been derived in the relevant literature for many specific problems, as well as for several mathematical programming subclasses. Despite the theoretical developments, it is not always clear that random projections are actually useful in solving mathematical programs in practice. In this paper we provide a computational assessment of the application of random projections to linear programming.","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"4 9","pages":"21:1-21:15"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91442412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Digraph k-Coloring Games: From Theory to Practice 有向图k-着色游戏:从理论到实践
Pub Date : 2022-01-01 DOI: 10.4230/LIPIcs.SEA.2022.20
Andrea D’ascenzo, Mattia D’emidio, M. Flammini, G. Monaco
We study digraph k -coloring games where agents are vertices of a directed unweighted graph and arcs represent agents’ mutual unidirectional idiosyncrasies or conflicts. Each agent can select one of k different colors, and her payoff in a given state is given by the number of outgoing neighbors with a different color. Such games model lots of strategic real-world scenarios and are related to several fundamental classes of anti-coordination games. Unfortunately, the problem of understanding whether an instance of the game admits a pure Nash equilibrium is NP-complete [33]. Therefore, in the last few years a relevant research focus has been that of designing polynomial time algorithms able to compute approximate Nash equilibria, i.e., states in which no agent, changing her strategy, can improve her payoff by some bounded multiplicative factor. The only two known algorithms in this respect are those in [14]. While they provide theoretical guarantees, their practical performance over real-world instances so far has not been investigated. In this paper, under the further motivation of the lack of practical approximation algorithms for the problem, we experimentally evaluate the above algorithms with the conclusion that, while they were suitably designed for achieving a bounded worst case behavior, they generally have a poor performance. Therefore, we next focus on classical best-response dynamics, and show that, despite of the fact that they might not always converge, they are very effective in practice. In particular, we provide a strong empirical evidence that they outperform existing methods, since surprisingly they quickly converge to exact Nash equilibria in almost all instances arising in practice. This also shows that, while this class of games is known to not always possess pure Nash equilibria, in almost all cases such equilibria exist and can be efficiently computed, even in a distributed uncoordinated way by a decentralized interaction of the agents. 2012 ACM Subject Classification Theory of computation → Algorithmic game theory and mechanism design; Theory of computation → Quality of equilibria; and analysis of algorithms;
我们研究了有向图k -着色博弈,其中代理是有向无权图的顶点,弧表示代理的相互单向特性或冲突。每个智能体可以从k种不同的颜色中选择一种,在给定状态下,她的收益由拥有不同颜色的邻居的数量给出。这类游戏模拟了许多现实世界的战略场景,并与若干基本类型的反协调游戏相关。不幸的是,理解博弈实例是否承认纯纳什均衡的问题是np完全的[33]。因此,在过去的几年里,一个相关的研究焦点是设计多项式时间算法来计算近似纳什均衡,即在这种状态下,没有智能体改变其策略,可以通过一些有界的乘法因子来提高其收益。在这方面仅有的两种已知算法是[14]中的算法。虽然它们提供了理论上的保证,但到目前为止,它们在现实世界实例中的实际性能尚未得到研究。在本文中,在缺乏实际逼近算法的进一步动机下,我们实验评估了上述算法,得出结论,尽管它们适合于实现有界的最坏情况行为,但它们通常性能较差。因此,我们接下来将关注经典的最佳响应动力学,并表明,尽管它们可能并不总是收敛,但它们在实践中是非常有效的。特别是,我们提供了强有力的经验证据,证明它们优于现有的方法,因为令人惊讶的是,它们几乎在实践中出现的所有情况下都能迅速收敛到精确的纳什均衡。这也表明,虽然这类博弈并不总是具有纯粹的纳什均衡,但在几乎所有情况下,这种均衡都存在,并且可以有效地计算,即使是通过分散的代理交互以分布式不协调的方式。2012 ACM学科分类:计算理论→算法博弈论与机制设计;计算理论→均衡质量;算法分析;
{"title":"Digraph k-Coloring Games: From Theory to Practice","authors":"Andrea D’ascenzo, Mattia D’emidio, M. Flammini, G. Monaco","doi":"10.4230/LIPIcs.SEA.2022.20","DOIUrl":"https://doi.org/10.4230/LIPIcs.SEA.2022.20","url":null,"abstract":"We study digraph k -coloring games where agents are vertices of a directed unweighted graph and arcs represent agents’ mutual unidirectional idiosyncrasies or conflicts. Each agent can select one of k different colors, and her payoff in a given state is given by the number of outgoing neighbors with a different color. Such games model lots of strategic real-world scenarios and are related to several fundamental classes of anti-coordination games. Unfortunately, the problem of understanding whether an instance of the game admits a pure Nash equilibrium is NP-complete [33]. Therefore, in the last few years a relevant research focus has been that of designing polynomial time algorithms able to compute approximate Nash equilibria, i.e., states in which no agent, changing her strategy, can improve her payoff by some bounded multiplicative factor. The only two known algorithms in this respect are those in [14]. While they provide theoretical guarantees, their practical performance over real-world instances so far has not been investigated. In this paper, under the further motivation of the lack of practical approximation algorithms for the problem, we experimentally evaluate the above algorithms with the conclusion that, while they were suitably designed for achieving a bounded worst case behavior, they generally have a poor performance. Therefore, we next focus on classical best-response dynamics, and show that, despite of the fact that they might not always converge, they are very effective in practice. In particular, we provide a strong empirical evidence that they outperform existing methods, since surprisingly they quickly converge to exact Nash equilibria in almost all instances arising in practice. This also shows that, while this class of games is known to not always possess pure Nash equilibria, in almost all cases such equilibria exist and can be efficiently computed, even in a distributed uncoordinated way by a decentralized interaction of the agents. 2012 ACM Subject Classification Theory of computation → Algorithmic game theory and mechanism design; Theory of computation → Quality of equilibria; and analysis of algorithms;","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"39 1","pages":"20:1-20:18"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84827653","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Parallel Framework for Approximate Max-Dicut in Partitionable Graphs 可分图中近似最大分割的并行框架
Pub Date : 2022-01-01 DOI: 10.4230/LIPIcs.SEA.2022.10
Nico Bertram, J. Ellert, J. Fischer
Computing a maximum cut in undirected and weighted graphs is a well studied problem and has many practical solutions that also scale well in shared memory (despite its NP-completeness). For its counterpart in directed graphs, however, we are not aware of practical solutions that also utilize parallelism. We engineer a framework that computes a high quality approximate cut in directed and weighted graphs by using a graph partitioning approach. The general idea is to partition a graph into k subgraphs using a parallel partitioning algorithm of our choice (the first ingredient of our framework). Then, for each subgraph in parallel, we compute a cut using any polynomial time approximation algorithm (the second ingredient). In a final step, we merge the locally computed solutions using a high-quality or exact parallel Max-Dicut algorithm (the third ingredient). On graphs that can be partitioned well, the quality of the computed cut is significantly better than the best cut achieved by any linear time algorithm. This is particularly relevant for large graphs, where linear time algorithms used to be the only feasible option.
在无向和加权图中计算最大割是一个研究得很好的问题,并且有许多实用的解决方案,它们也可以很好地扩展到共享内存中(尽管它具有np完备性)。然而,对于有向图中的对应对象,我们还不知道也利用并行性的实际解决方案。我们设计了一个框架,通过使用图划分方法在有向图和加权图中计算高质量的近似切。总体思路是使用我们选择的并行划分算法(我们框架的第一个组成部分)将一个图划分为k个子图。然后,对于并行的每个子图,我们使用任何多项式时间近似算法(第二个成分)计算切割。在最后一步中,我们使用高质量或精确并行的Max-Dicut算法(第三个成分)合并局部计算的解。在可以很好分割的图上,计算得到的切的质量明显优于任何线性时间算法得到的最佳切。这对于大型图来说尤其重要,因为线性时间算法曾经是唯一可行的选择。
{"title":"A Parallel Framework for Approximate Max-Dicut in Partitionable Graphs","authors":"Nico Bertram, J. Ellert, J. Fischer","doi":"10.4230/LIPIcs.SEA.2022.10","DOIUrl":"https://doi.org/10.4230/LIPIcs.SEA.2022.10","url":null,"abstract":"Computing a maximum cut in undirected and weighted graphs is a well studied problem and has many practical solutions that also scale well in shared memory (despite its NP-completeness). For its counterpart in directed graphs, however, we are not aware of practical solutions that also utilize parallelism. We engineer a framework that computes a high quality approximate cut in directed and weighted graphs by using a graph partitioning approach. The general idea is to partition a graph into k subgraphs using a parallel partitioning algorithm of our choice (the first ingredient of our framework). Then, for each subgraph in parallel, we compute a cut using any polynomial time approximation algorithm (the second ingredient). In a final step, we merge the locally computed solutions using a high-quality or exact parallel Max-Dicut algorithm (the third ingredient). On graphs that can be partitioned well, the quality of the computed cut is significantly better than the best cut achieved by any linear time algorithm. This is particularly relevant for large graphs, where linear time algorithms used to be the only feasible option.","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"12 1","pages":"10:1-10:15"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86540919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Relating Real and Synthetic Social Networks Through Centrality Measures 通过中心性度量来关联真实的和合成的社会网络
Pub Date : 2022-01-01 DOI: 10.4230/LIPIcs.SEA.2022.7
M. Blesa, Mihail Eduard Popa, M. Serna
We perform here a comparative study on the behaviour of real and synthetic social networks with respect to a selection of nine centrality measures. Some of them are topology based (degree, closeness, betweenness), while others consider the relevance of the actors within the network (Katz, PageRank) or their ability to spread influence through it (Independent Cascade rank, Linear Threshold Rank). We run different experiments on synthetic social networks, with 1K, 10K, and 100K nodes, generated according to the Gaussian Random partition model, the stochastic block model, the LFR benchmark graph model and hyperbolic geometric graphs model. Some real social networks are also considered, with the aim of discovering how do they relate to the synthetic models in terms of centrality. Apart from usual statistical measures, we perform a correlation analysis between all the nine measures. Our results indicate that, in general, the correlation matrices of the different models scale nicely with size. Moreover, the correlation plots distinguish four categories that classify most of the real networks studied here. Those categories have a clear correspondence with particular configurations of the models for synthetic networks.
我们在这里执行的行为比较研究的真实和合成的社会网络相对于九中心性措施的选择。其中一些是基于拓扑的(度,亲密度,中间度),而另一些则考虑网络中参与者的相关性(Katz, PageRank)或他们通过网络传播影响的能力(独立级联排名,线性阈值排名)。我们根据高斯随机划分模型、随机块模型、LFR基准图模型和双曲几何图模型,在合成社交网络上进行了不同的实验,节点分别为1K、10K和100K。还考虑了一些真实的社交网络,目的是发现它们在中心性方面如何与合成模型相关联。除了通常的统计措施外,我们还对所有九个措施进行了相关性分析。我们的研究结果表明,在一般情况下,不同模型的相关矩阵与尺寸的比例很好。此外,相关图将本文研究的大多数真实网络划分为四类。这些类别与合成网络模型的特定配置有明确的对应关系。
{"title":"Relating Real and Synthetic Social Networks Through Centrality Measures","authors":"M. Blesa, Mihail Eduard Popa, M. Serna","doi":"10.4230/LIPIcs.SEA.2022.7","DOIUrl":"https://doi.org/10.4230/LIPIcs.SEA.2022.7","url":null,"abstract":"We perform here a comparative study on the behaviour of real and synthetic social networks with respect to a selection of nine centrality measures. Some of them are topology based (degree, closeness, betweenness), while others consider the relevance of the actors within the network (Katz, PageRank) or their ability to spread influence through it (Independent Cascade rank, Linear Threshold Rank). We run different experiments on synthetic social networks, with 1K, 10K, and 100K nodes, generated according to the Gaussian Random partition model, the stochastic block model, the LFR benchmark graph model and hyperbolic geometric graphs model. Some real social networks are also considered, with the aim of discovering how do they relate to the synthetic models in terms of centrality. Apart from usual statistical measures, we perform a correlation analysis between all the nine measures. Our results indicate that, in general, the correlation matrices of the different models scale nicely with size. Moreover, the correlation plots distinguish four categories that classify most of the real networks studied here. Those categories have a clear correspondence with particular configurations of the models for synthetic networks.","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"39 1","pages":"7:1-7:21"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81110596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Experimental Study of Algorithms for Packing Arborescences 树形排列填充算法的实验研究
Pub Date : 2022-01-01 DOI: 10.4230/LIPIcs.SEA.2022.14
L. Georgiadis, Dionysios Kefallinos, Anna Mpanti, Stavros D. Nikolopoulos
A classic result of Edmonds states that the maximum number of edge-disjoint arborescences of a directed graph G , rooted at a designated vertex s , equals the minimum cardinality c G ( s ) of an s -cut of G . This concept is related to the edge connectivity λ ( G ) of a strongly connected directed graph G , defined as the minimum number of edges whose deletion leaves a graph that is not strongly connected. In this paper, we address the question of how efficiently we can compute a maximum packing of edge-disjoint arborescences in practice, compared to the time required to determine the edge connectivity of a graph. To that end, we explore the design space of efficient algorithms for packing arborescences of a directed graph in practice and conduct a thorough empirical study to highlight the merits and weaknesses of each technique. In particular, we present an efficient implementation of Gabow’s arborescence packing algorithm and provide a simple but efficient heuristic that significantly improves its running time in practice. Project FANTA (eFficient Algorithms for NeTwork Analysis), number HFRI-FM17-431. H = G − A , and a complete ( k − 1)-intersection T for s on G . It uses the following key concept. An enlarging path consists of an edge e ∈ E + ( A ), and if e ∈ T , an augmenting path P for the ( k − 1)-intersection T − e on H − e . Gabow shows that if V ( A ) ̸ = V ( G ), then there is always an enlarging path.
Edmonds的一个经典结果表明,根于指定顶点s的有向图G的最大边不相交树形数等于G的s -cut的最小基数c G (s)。这个概念与强连接有向图G的边连通性λ (G)有关,定义为其删除后留下非强连接图的最小边数。在本文中,我们解决的问题是,与确定图的边缘连通性所需的时间相比,我们在实践中如何有效地计算边缘不相交树的最大包装。为此,我们在实践中探索了有向图树杈填充的有效算法的设计空间,并进行了彻底的实证研究,以突出每种技术的优点和缺点。特别地,我们提出了Gabow树形打包算法的有效实现,并提供了一个简单而有效的启发式算法,在实践中显著提高了其运行时间。网络分析的高效算法(FANTA)项目,编号hri - fm17 -431。H = G−A,以及s在G上的完全(k−1)交T。它使用以下关键概念。扩展路径由一条边e∈e + (A)组成,如果e∈T,则在H−e上的(k−1)交T−e的扩展路径P。Gabow表明,如果V (A) h = V (G),则总存在一条扩大路径。
{"title":"An Experimental Study of Algorithms for Packing Arborescences","authors":"L. Georgiadis, Dionysios Kefallinos, Anna Mpanti, Stavros D. Nikolopoulos","doi":"10.4230/LIPIcs.SEA.2022.14","DOIUrl":"https://doi.org/10.4230/LIPIcs.SEA.2022.14","url":null,"abstract":"A classic result of Edmonds states that the maximum number of edge-disjoint arborescences of a directed graph G , rooted at a designated vertex s , equals the minimum cardinality c G ( s ) of an s -cut of G . This concept is related to the edge connectivity λ ( G ) of a strongly connected directed graph G , defined as the minimum number of edges whose deletion leaves a graph that is not strongly connected. In this paper, we address the question of how efficiently we can compute a maximum packing of edge-disjoint arborescences in practice, compared to the time required to determine the edge connectivity of a graph. To that end, we explore the design space of efficient algorithms for packing arborescences of a directed graph in practice and conduct a thorough empirical study to highlight the merits and weaknesses of each technique. In particular, we present an efficient implementation of Gabow’s arborescence packing algorithm and provide a simple but efficient heuristic that significantly improves its running time in practice. Project FANTA (eFficient Algorithms for NeTwork Analysis), number HFRI-FM17-431. H = G − A , and a complete ( k − 1)-intersection T for s on G . It uses the following key concept. An enlarging path consists of an edge e ∈ E + ( A ), and if e ∈ T , an augmenting path P for the ( k − 1)-intersection T − e on H − e . Gabow shows that if V ( A ) ̸ = V ( G ), then there is always an enlarging path.","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"1 1","pages":"14:1-14:16"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75918320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Solving and Generating Nagareru Puzzles 解决和生成Nagareru谜题
Pub Date : 2022-01-01 DOI: 10.4230/LIPIcs.SEA.2022.2
Masakazu Ishihata, Fumiya Tokumasu
Solving paper-and-pencil puzzles is fun for people, and their analysis is also an essential issue in computational complexity theory. There are some practically efficient solvers for some NP-complete puzzles; however, the automatic generation of interesting puzzle instances still stands out as a complex problem because it requires checking whether the generated instance has a unique solution. In this paper, we focus on a puzzle called Nagareru and propose two methods: one is for implicitly enumerating all the solutions of its instance, and the other is for efficiently generating an instance with a unique solution. The former constructs a ZDD that implicitly represents all the solutions. The latter employs the ZDD-based solver as a building block to check the uniqueness of the solution of generated instances. We experimentally showed that the ZDD-based solver was drastically faster than a CSP-based solver, and our generation method created an interesting instance in a reasonable time.
解决纸笔难题对人们来说是一种乐趣,而它们的分析也是计算复杂性理论中的一个重要问题。对于某些np完全谜题,确实存在一些有效的解法;然而,自动生成有趣的谜题实例仍然是一个复杂的问题,因为它需要检查生成的实例是否具有唯一的解决方案。本文研究了一个名为Nagareru的谜题,并提出了两种方法:一种是隐式枚举其实例的所有解,另一种是有效地生成具有唯一解的实例。前者构造一个隐式表示所有解决方案的ZDD。后者使用基于zdd的求解器作为构建块来检查生成实例的解决方案的唯一性。我们通过实验证明,基于zdd的求解器比基于csp的求解器要快得多,并且我们的生成方法在合理的时间内创建了一个有趣的实例。
{"title":"Solving and Generating Nagareru Puzzles","authors":"Masakazu Ishihata, Fumiya Tokumasu","doi":"10.4230/LIPIcs.SEA.2022.2","DOIUrl":"https://doi.org/10.4230/LIPIcs.SEA.2022.2","url":null,"abstract":"Solving paper-and-pencil puzzles is fun for people, and their analysis is also an essential issue in computational complexity theory. There are some practically efficient solvers for some NP-complete puzzles; however, the automatic generation of interesting puzzle instances still stands out as a complex problem because it requires checking whether the generated instance has a unique solution. In this paper, we focus on a puzzle called Nagareru and propose two methods: one is for implicitly enumerating all the solutions of its instance, and the other is for efficiently generating an instance with a unique solution. The former constructs a ZDD that implicitly represents all the solutions. The latter employs the ZDD-based solver as a building block to check the uniqueness of the solution of generated instances. We experimentally showed that the ZDD-based solver was drastically faster than a CSP-based solver, and our generation method created an interesting instance in a reasonable time.","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"60 1","pages":"2:1-2:17"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80510423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Exact Learning Algorithms for Road Networks and Other Graphs with Bounded Clustering Degrees 具有有界聚类度的道路网络和其他图的高效精确学习算法
Pub Date : 2022-01-01 DOI: 10.4230/LIPIcs.SEA.2022.9
Ramtin Afshar, M. Goodrich, Evrim Ozel
The completeness of road network data is significant in the quality of various routing services and applications. We introduce an efficient randomized algorithm for exact learning of road networks using simple distance queries, which can find missing roads and improve the quality of routing services. The efficiency of our algorithm depends on a cluster degree parameter, d max , which is an upper bound on the degrees of vertex clusters defined during our algorithm. Unfortunately, we leave open the problem of theoretically bounding d max , although we conjecture that d max is small for road networks and other similar types of graphs. We support this conjecture by experimentally evaluating our algorithm on road network data for the U.S. and 5 European countries of various sizes. This analysis provides experimental evidence that our algorithm issues a quasilinear number of queries in expectation for road networks and similar graphs. of → Graph analysis; Theory of → Random network models;
路网数据的完整性对各种路由服务和应用的质量具有重要意义。我们引入了一种高效的随机算法,利用简单的距离查询来精确学习道路网络,可以找到缺失的道路,提高路由服务的质量。算法的效率取决于聚类度参数d max,它是算法中定义的顶点聚类度的上界。不幸的是,尽管我们推测d max对于道路网络和其他类似类型的图来说很小,但我们仍然没有解决理论上d max的边界问题。我们通过在美国和5个不同规模的欧洲国家的道路网络数据上实验评估我们的算法来支持这一猜想。该分析提供了实验证据,表明我们的算法在道路网络和类似图的期望中发出拟线性查询数。of→图分析;→随机网络模型理论;
{"title":"Efficient Exact Learning Algorithms for Road Networks and Other Graphs with Bounded Clustering Degrees","authors":"Ramtin Afshar, M. Goodrich, Evrim Ozel","doi":"10.4230/LIPIcs.SEA.2022.9","DOIUrl":"https://doi.org/10.4230/LIPIcs.SEA.2022.9","url":null,"abstract":"The completeness of road network data is significant in the quality of various routing services and applications. We introduce an efficient randomized algorithm for exact learning of road networks using simple distance queries, which can find missing roads and improve the quality of routing services. The efficiency of our algorithm depends on a cluster degree parameter, d max , which is an upper bound on the degrees of vertex clusters defined during our algorithm. Unfortunately, we leave open the problem of theoretically bounding d max , although we conjecture that d max is small for road networks and other similar types of graphs. We support this conjecture by experimentally evaluating our algorithm on road network data for the U.S. and 5 European countries of various sizes. This analysis provides experimental evidence that our algorithm issues a quasilinear number of queries in expectation for road networks and similar graphs. of → Graph analysis; Theory of → Random network models;","PeriodicalId":9448,"journal":{"name":"Bulletin of the Society of Sea Water Science, Japan","volume":"102 1","pages":"9:1-9:18"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76769125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Bulletin of the Society of Sea Water Science, Japan
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1