首页 > 最新文献

2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)最新文献

英文 中文
Coverability Analysis of Time Basic Petri Nets with Non-Urgent Behavior 非紧急行为时基Petri网的可覆盖性分析
Matteo Camilli, C. Bellettini, L. Capra, Mattia Monga
Time Basic Petri nets are an expressive extension of Petri nets, suitable to model real-time systems. This paper introduces a coverability analysis technique to cope with structurally unbounded Time Basic Petri net models exhibiting non-urgent behavior: i.e., models in which transitions may choose to do not fire and let time pass, even if this could lead to transition disabling. The approach we present exploits the identification of anonymous temporal information, that is the possibility of erasing timestamps associated with specific tokens without compromising the correctness of model's temporal evolution. In particular, we extend the classical Karp-Miller coverability algorithm in two ways: first, we adapt the acceleration function to deal with symbolic states and to identify unboundedness due to time anonymous tokens, second, we employ an aggressive pruning strategy to remove included/covered portions of the reachability tree during exploration.
时间基础Petri网是Petri网的表达扩展,适用于实时系统建模。本文介绍了一种可覆盖性分析技术,以应对结构无界的时间基本Petri网模型表现出非紧急行为:即,模型中的转换可能选择不触发并让时间过去,即使这可能导致转换失效。我们提出的方法利用了匿名时间信息的识别,即在不影响模型时间演化的正确性的情况下删除与特定令牌相关的时间戳的可能性。特别是,我们从两个方面扩展了经典的Karp-Miller可覆盖性算法:首先,我们调整加速函数来处理符号状态并识别由于时间匿名令牌而导致的无界性;其次,我们采用积极的修剪策略来删除在探索期间可达性树的包含/覆盖部分。
{"title":"Coverability Analysis of Time Basic Petri Nets with Non-Urgent Behavior","authors":"Matteo Camilli, C. Bellettini, L. Capra, Mattia Monga","doi":"10.1109/SYNASC.2016.036","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.036","url":null,"abstract":"Time Basic Petri nets are an expressive extension of Petri nets, suitable to model real-time systems. This paper introduces a coverability analysis technique to cope with structurally unbounded Time Basic Petri net models exhibiting non-urgent behavior: i.e., models in which transitions may choose to do not fire and let time pass, even if this could lead to transition disabling. The approach we present exploits the identification of anonymous temporal information, that is the possibility of erasing timestamps associated with specific tokens without compromising the correctness of model's temporal evolution. In particular, we extend the classical Karp-Miller coverability algorithm in two ways: first, we adapt the acceleration function to deal with symbolic states and to identify unboundedness due to time anonymous tokens, second, we employ an aggressive pruning strategy to remove included/covered portions of the reachability tree during exploration.","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130578353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Partial Finitely Generated Bi-Ideals 部分有限生成双理想
Raivis Bēts, J. Buls
Partial words have been studied by Blanchet-Sadri et al., but bi-ideals or reccurrent words have been studied for centuries by many researchers. This paper gives a solution for some problems for partial reccurrent words. This paper gives an algorithm for a given finitely generated bi-ideal, how to construct a new basis of ultimately finitely generated bi-ideal, which generates the same given bi-ideal. The paper states that it is always possible to find a basis for a given finitely generated bi-ideal. The main results of this paper are presented in third section. At first, we show that if two irreduciable bi-ideals are different, they will differ in infinitely many places. This led to the statement that it is possible to fill the finite number of holes for a given finitely generated bi-ideal, but a counterexample shows that it is not possible for infinitely many holes.
Blanchet-Sadri等人对部分词进行了研究,但双理想词或循环词已经被许多研究者研究了几个世纪。针对部分重复词的一些问题,给出了一种解决方法。本文给出了给定有限生成双理想的一种算法,即如何构造最终有限生成双理想的一组新基,从而生成同一个给定双理想。本文指出,对于给定的有限生成双理想,总有可能找到一个基。第三部分给出了本文的主要研究结果。首先,我们证明了如果两个不可约双理想是不同的,那么它们在无限多的地方是不同的。这导致了对于给定的有限生成的双理想,有可能填充有限数量的空穴,但反例表明对于无限多的空穴,这是不可能的。
{"title":"Partial Finitely Generated Bi-Ideals","authors":"Raivis Bēts, J. Buls","doi":"10.1109/SYNASC.2016.065","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.065","url":null,"abstract":"Partial words have been studied by Blanchet-Sadri et al., but bi-ideals or reccurrent words have been studied for centuries by many researchers. This paper gives a solution for some problems for partial reccurrent words. This paper gives an algorithm for a given finitely generated bi-ideal, how to construct a new basis of ultimately finitely generated bi-ideal, which generates the same given bi-ideal. The paper states that it is always possible to find a basis for a given finitely generated bi-ideal. The main results of this paper are presented in third section. At first, we show that if two irreduciable bi-ideals are different, they will differ in infinitely many places. This led to the statement that it is possible to fill the finite number of holes for a given finitely generated bi-ideal, but a counterexample shows that it is not possible for infinitely many holes.","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132009252","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identifying DGA-Based Botnets Using Network Anomaly Detection 利用网络异常检测识别基于dga的僵尸网络
Dragos Gavrilut, George Popoiu, Razvan Benchea
Nowadays, the attacks are no longer performed from a single computer but from thousands, sometimes millions of systems that are located all over the globe and are grouped in a network called botnet. The most widely used technique to control a botnet is to try to connect to many domain names, generated according to an algorithm called domain generating algorithm (DGA). In this paper we present different algorithms that can determine if a computer is part of a botnet by looking at its network traffic. Since in some cases the network traffic is impossible to be shared due to privacy reasons we also analyze the case where just limited information can be provided (such as a netflow log). The algorithms presented here were obtained after reverse engineering and analyzing the DGA of 18 different botnets including some that were taken down (such as Cryptolocker) and ones that are still alive and thriving (such as PushDo, Tinba, Nivdort, DirtyLocker, Dobot, Patriot, Ramdo, Virut, Ramnit and many more).
如今,攻击不再是从一台计算机上进行的,而是来自全球各地的数千台,有时甚至数百万台系统,这些系统被分组在一个称为僵尸网络的网络中。控制僵尸网络最常用的技术是尝试连接到许多域名,这些域名是根据一种称为域生成算法(DGA)的算法生成的。在本文中,我们提出了不同的算法,可以通过查看其网络流量来确定计算机是否是僵尸网络的一部分。由于在某些情况下,由于隐私原因,网络流量是不可能共享的,我们还分析了只能提供有限信息的情况(如netflow日志)。本文介绍的算法是在逆向工程和分析18个不同僵尸网络的DGA后获得的,其中包括一些被关闭的僵尸网络(如Cryptolocker)和那些仍然存在并蓬勃发展的僵尸网络(如PushDo, Tinba, nivort, DirtyLocker, Dobot, Patriot, Ramdo, Virut, Ramnit等等)。
{"title":"Identifying DGA-Based Botnets Using Network Anomaly Detection","authors":"Dragos Gavrilut, George Popoiu, Razvan Benchea","doi":"10.1109/SYNASC.2016.053","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.053","url":null,"abstract":"Nowadays, the attacks are no longer performed from a single computer but from thousands, sometimes millions of systems that are located all over the globe and are grouped in a network called botnet. The most widely used technique to control a botnet is to try to connect to many domain names, generated according to an algorithm called domain generating algorithm (DGA). In this paper we present different algorithms that can determine if a computer is part of a botnet by looking at its network traffic. Since in some cases the network traffic is impossible to be shared due to privacy reasons we also analyze the case where just limited information can be provided (such as a netflow log). The algorithms presented here were obtained after reverse engineering and analyzing the DGA of 18 different botnets including some that were taken down (such as Cryptolocker) and ones that are still alive and thriving (such as PushDo, Tinba, Nivdort, DirtyLocker, Dobot, Patriot, Ramdo, Virut, Ramnit and many more).","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123282863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Convergence Acceleration of Iterative Methods for Inverting Real Matrices Using Frobenius Norm Minimization 利用Frobenius范数最小化求实矩阵逆的迭代方法的收敛加速
Ajinkya Borle, S. Lomonaco
The Schulz-type methods for computing generalizedmatrix inverses are a family of iterative methods that are popular for their high order of convergence (≥ 2). We propose two new scaled acceleration techniques for such type of iterative methods for real matrices (based on Frobenius norm minimization) andlay out efficient algorithms to implement these techniques. Testresults show one of our techniques to be most effective for densematrices but also works for sparse cases as well.
计算广义矩阵逆的schulz型方法是一系列迭代方法,它们因其高收敛阶(≥2)而广受欢迎。我们提出了两种新的基于Frobenius范数最小化的实际矩阵的此类迭代方法的缩放加速技术,并提出了实现这些技术的有效算法。测试结果表明,我们的一种技术对密集矩阵最有效,但也适用于稀疏情况。
{"title":"Convergence Acceleration of Iterative Methods for Inverting Real Matrices Using Frobenius Norm Minimization","authors":"Ajinkya Borle, S. Lomonaco","doi":"10.1109/SYNASC.2016.031","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.031","url":null,"abstract":"The Schulz-type methods for computing generalizedmatrix inverses are a family of iterative methods that are popular for their high order of convergence (≥ 2). We propose two new scaled acceleration techniques for such type of iterative methods for real matrices (based on Frobenius norm minimization) andlay out efficient algorithms to implement these techniques. Testresults show one of our techniques to be most effective for densematrices but also works for sparse cases as well.","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133147384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Comparing Different Term Weighting Schemas for Topic Modeling 主题建模中不同词权重模式的比较
Ciprian-Octavian Truică, F. Rădulescu, A. Boicea
Topic Modeling is a type of statistical model that tries to determine the topics present in a corpus of documents. The accuracy measures applied to clustering algorithm can also be used to assess the accuracy of topic modeling algorithms because determining topics for documents is similar with clustering them. This paper presents an experimental validation regarding the accuracy of Latent Dirichlet Allocation in comparison with Non-Negative Matrix Factorization and K-Means. The experiments use different weighting schemas when constructing the document-term matrix to determine if the accuracy of the algorithm improves. Two well known, already labeled text corpora are used for testing. The Purity and Adjusted Rand Index are used to evaluate the accuracy. Also, a time performance comparison regarding the run-time of these algorithms is presented.
主题建模是一种统计模型,它试图确定文档语料库中存在的主题。应用于聚类算法的精度度量也可用于评估主题建模算法的准确性,因为确定文档的主题与聚类它们类似。本文通过实验验证了潜狄利克雷分配方法与非负矩阵分解和K-Means方法的准确性。实验采用不同的加权模式来构建文档项矩阵,以确定算法的准确率是否有所提高。两个众所周知的,已经标记的文本语料库用于测试。使用纯度和调整后的兰德指数来评估准确性。同时,对这些算法在运行时的时间性能进行了比较。
{"title":"Comparing Different Term Weighting Schemas for Topic Modeling","authors":"Ciprian-Octavian Truică, F. Rădulescu, A. Boicea","doi":"10.1109/SYNASC.2016.055","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.055","url":null,"abstract":"Topic Modeling is a type of statistical model that tries to determine the topics present in a corpus of documents. The accuracy measures applied to clustering algorithm can also be used to assess the accuracy of topic modeling algorithms because determining topics for documents is similar with clustering them. This paper presents an experimental validation regarding the accuracy of Latent Dirichlet Allocation in comparison with Non-Negative Matrix Factorization and K-Means. The experiments use different weighting schemas when constructing the document-term matrix to determine if the accuracy of the algorithm improves. Two well known, already labeled text corpora are used for testing. The Purity and Adjusted Rand Index are used to evaluate the accuracy. Also, a time performance comparison regarding the run-time of these algorithms is presented.","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115198333","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
A Numerical Method for Analyzing the Stability of Bi-Parametric Biological Systems 双参数生物系统稳定性分析的一种数值方法
Changbo Chen, Wenyuan Wu
For a biological system modeled by a continuous dynamical system defined by rational functions with two parameters, we propose a numerical method to compute the fold and Hopf bifurcation boundaries of the system restricted in a finite region in the parametric space under certain assumptions. The bifurcation boundaries divide their complement in the region into connected subsets, called cells, such that above each of them the number of equilibria is constant and the stability of each equilibrium remains unchanged. The boundaries are generated by first tracing the fold and Hopf bifurcation curves in a higher dimensional space and then projecting them onto the parameter plane. One advantage of this method is that it can exploit global information of real varieties and generate complete boundaries based on homotopy continuation methods and critical point techniques. The bistability properties of several biological systems are successfully analyzed by our method.
对于由两个参数有理函数定义的连续动力系统建模的生物系统,在一定的假设条件下,给出了在参数空间有限区域内计算该系统的折叠边界和Hopf分岔边界的数值方法。分支边界将其在区域内的补体划分为相互连接的子集,称为细胞,使得每个子集以上的平衡点数量不变,并且每个平衡点的稳定性保持不变。首先在高维空间中跟踪折叠和Hopf分岔曲线,然后将其投影到参数平面上,从而生成边界。该方法的一个优点是利用实变量的全局信息,利用同伦延拓方法和临界点技术生成完全边界。用我们的方法成功地分析了几种生物系统的双稳性。
{"title":"A Numerical Method for Analyzing the Stability of Bi-Parametric Biological Systems","authors":"Changbo Chen, Wenyuan Wu","doi":"10.1109/SYNASC.2016.026","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.026","url":null,"abstract":"For a biological system modeled by a continuous dynamical system defined by rational functions with two parameters, we propose a numerical method to compute the fold and Hopf bifurcation boundaries of the system restricted in a finite region in the parametric space under certain assumptions. The bifurcation boundaries divide their complement in the region into connected subsets, called cells, such that above each of them the number of equilibria is constant and the stability of each equilibrium remains unchanged. The boundaries are generated by first tracing the fold and Hopf bifurcation curves in a higher dimensional space and then projecting them onto the parameter plane. One advantage of this method is that it can exploit global information of real varieties and generate complete boundaries based on homotopy continuation methods and critical point techniques. The bistability properties of several biological systems are successfully analyzed by our method.","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"14 12","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114111381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
In Search of the Optimal Set of Indicators when Classifying Histopathological Images 组织病理图像分类中最优指标集的寻找
C. Stoean
There is currently a large amount of histopathological images due to the intensive prevention screening programs worldwide. This fact overloads the pathologists' tasks. Hence, there is a connected high need for a quantitative image-based evaluation of digital pathology slides. The current work extracts 76 numerical features from 357 histopathological images and focuses on the selection of the most valuable features that conducts to a smaller data set on which a SVM classifier achieves a better prediction. The gain in accuracy is of over 4% more than in the situation when the entire data set was used. The paper also indicates a subset of the attributes that proved to be the most informative with respect to 4 feature selection approaches.
目前有大量的组织病理学图像由于密集的预防筛查计划在世界范围内。这一事实加重了病理学家的工作负担。因此,对数字病理切片的定量图像评估有很高的需求。目前的工作从357张组织病理学图像中提取了76个数字特征,并着重于选择最有价值的特征,这些特征可以用于更小的数据集,在这些数据集上SVM分类器可以实现更好的预测。与使用整个数据集的情况相比,准确度提高了4%以上。本文还指出了在4种特征选择方法中被证明是最具信息量的属性子集。
{"title":"In Search of the Optimal Set of Indicators when Classifying Histopathological Images","authors":"C. Stoean","doi":"10.1109/SYNASC.2016.074","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.074","url":null,"abstract":"There is currently a large amount of histopathological images due to the intensive prevention screening programs worldwide. This fact overloads the pathologists' tasks. Hence, there is a connected high need for a quantitative image-based evaluation of digital pathology slides. The current work extracts 76 numerical features from 357 histopathological images and focuses on the selection of the most valuable features that conducts to a smaller data set on which a SVM classifier achieves a better prediction. The gain in accuracy is of over 4% more than in the situation when the entire data set was used. The paper also indicates a subset of the attributes that proved to be the most informative with respect to 4 feature selection approaches.","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126172062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Efficient and Robust Median-of-Means Algorithms for Location and Regression 定位与回归的高效鲁棒中值算法
Alexander Kogler, Patrick Traxler
We consider the computational problem to learn models from data that are possibly contaminated with outliers. We design and analyze algorithms for robust location and robust linear regression. Such algorithms are essential for solving central problems of robust statistics and outlier detection. We show that our algorithms, which are based on a novel extension of the Median-of-Means method by employing the discrete geometric median, are accurate, efficient and robust against many outliers in the data. The discrete geometric median has many desirable characteristics such as it works for general metric spaces and preserves combinatorial and statistical properties. Furthermore, there is an exact and efficient algorithm to compute it, and an even faster approximation algorithm. We present theoretical and experimental results. In particular, we emphasize the generality of Median-of-Means and its ability to speedup and parallelize algorithms which additionally are accurate and robust against many outliers in the data.
我们考虑从可能被异常值污染的数据中学习模型的计算问题。我们设计并分析了鲁棒定位和鲁棒线性回归的算法。这些算法对于解决鲁棒统计和离群值检测的核心问题至关重要。我们的算法基于采用离散几何中位数的中位数方法的新扩展,对数据中的许多异常值具有准确,高效和鲁棒性。离散几何中位数具有许多理想的特性,例如它适用于一般度量空间,并保留了组合和统计性质。此外,有一个精确而有效的算法来计算它,甚至更快的近似算法。我们提出了理论和实验结果。特别地,我们强调了中位数的通用性及其加速和并行化算法的能力,这些算法对数据中的许多异常值具有准确和鲁棒性。
{"title":"Efficient and Robust Median-of-Means Algorithms for Location and Regression","authors":"Alexander Kogler, Patrick Traxler","doi":"10.1109/SYNASC.2016.041","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.041","url":null,"abstract":"We consider the computational problem to learn models from data that are possibly contaminated with outliers. We design and analyze algorithms for robust location and robust linear regression. Such algorithms are essential for solving central problems of robust statistics and outlier detection. We show that our algorithms, which are based on a novel extension of the Median-of-Means method by employing the discrete geometric median, are accurate, efficient and robust against many outliers in the data. The discrete geometric median has many desirable characteristics such as it works for general metric spaces and preserves combinatorial and statistical properties. Furthermore, there is an exact and efficient algorithm to compute it, and an even faster approximation algorithm. We present theoretical and experimental results. In particular, we emphasize the generality of Median-of-Means and its ability to speedup and parallelize algorithms which additionally are accurate and robust against many outliers in the data.","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126465012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Towards Lexicalization of DBpedia Ontology with Unsupervised Learning and Semantic Role Labeling 基于无监督学习和语义角色标注的DBpedia本体词汇化研究
A. Marginean, Kando Eniko
Filling the gap between natural language expressions and ontology concepts or properties is the new trend in Semantic Web. Ontology lexicalization introduces a new layer of lexical information for ontology properties and concepts. We propose a method based on unsupervised learning for the extraction of the potential lexical expressions of DBpedia propertiesfrom Wikipedia text corpus. It is a resource-driven approach that comprises three main steps. The first step consists of the extraction of DBpedia triples for the aimed property followed by the extraction of Wikipedia articles describing the resources from these triples. In the second step, sentences mostly related to the property are extracted from the articles and they are analyzed with a Semantic Role Labeler resulting in a set of SRL annotated trees. In the last step, clusters of expressions are built using spectral clustering based on the distances between the SRL trees. The clusters with the least variance are considered to be relevant for the lexical expressions of the property.
填补自然语言表达式与本体概念或属性之间的空白是语义Web的新趋势。本体词汇化为本体属性和概念引入了一个新的词汇信息层。本文提出了一种基于无监督学习的方法,用于从维基百科文本语料库中提取DBpedia属性的潜在词汇表达式。这是一种资源驱动的方法,包括三个主要步骤。第一步包括为目标属性提取DBpedia三元组,然后从这些三元组中提取描述资源的Wikipedia文章。在第二步中,从文章中提取与属性相关的句子,并使用语义角色标签器对其进行分析,从而生成一组SRL注释树。在最后一步中,基于SRL树之间的距离,使用光谱聚类构建表达簇。方差最小的聚类被认为与属性的词法表达式相关。
{"title":"Towards Lexicalization of DBpedia Ontology with Unsupervised Learning and Semantic Role Labeling","authors":"A. Marginean, Kando Eniko","doi":"10.1109/SYNASC.2016.048","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.048","url":null,"abstract":"Filling the gap between natural language expressions and ontology concepts or properties is the new trend in Semantic Web. Ontology lexicalization introduces a new layer of lexical information for ontology properties and concepts. We propose a method based on unsupervised learning for the extraction of the potential lexical expressions of DBpedia propertiesfrom Wikipedia text corpus. It is a resource-driven approach that comprises three main steps. The first step consists of the extraction of DBpedia triples for the aimed property followed by the extraction of Wikipedia articles describing the resources from these triples. In the second step, sentences mostly related to the property are extracted from the articles and they are analyzed with a Semantic Role Labeler resulting in a set of SRL annotated trees. In the last step, clusters of expressions are built using spectral clustering based on the distances between the SRL trees. The clusters with the least variance are considered to be relevant for the lexical expressions of the property.","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116985009","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Identifiability for Gauge Regularizations and Algorithms for Block-Sparse Synthesis in Compressive Sensing 压缩感知中规范正则化的可辨识性及块稀疏合成算法
F. Turcu, C. Dossal, Marc Nicodeme
In the paper we give a characterization of identifiability for regularizations with gauges of compact convexes. This extends the classic identifiability results from the standard l1-regularization framework in compressive sensing. We show that the standard dual certificate techniques can no longer work by themselves ouside the polytope case. We then apply the general characterization to the caseof block-sparse regularizations and obtain an identification algorithm based on a combination of the standard duality and a convex-projection technique.
本文给出了紧凸规正则化的可辨识性的一个表征。这扩展了压缩感知中标准11 -正则化框架的经典可识别性结果。我们证明了标准的双证书技术在多面体情况之外不能再单独工作。然后,我们将一般表征应用于块稀疏正则化的情况,并获得了一种基于标准对偶和凸投影技术相结合的识别算法。
{"title":"Identifiability for Gauge Regularizations and Algorithms for Block-Sparse Synthesis in Compressive Sensing","authors":"F. Turcu, C. Dossal, Marc Nicodeme","doi":"10.1109/SYNASC.2016.029","DOIUrl":"https://doi.org/10.1109/SYNASC.2016.029","url":null,"abstract":"In the paper we give a characterization of identifiability for regularizations with gauges of compact convexes. This extends the classic identifiability results from the standard l1-regularization framework in compressive sensing. We show that the standard dual certificate techniques can no longer work by themselves ouside the polytope case. We then apply the general characterization to the caseof block-sparse regularizations and obtain an identification algorithm based on a combination of the standard duality and a convex-projection technique.","PeriodicalId":268635,"journal":{"name":"2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)","volume":"139 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127534190","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2016 18th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1