首页 > 最新文献

2011 IEEE 23rd International Conference on Tools with Artificial Intelligence最新文献

英文 中文
Classification of Hyperspectral Imagery Using GPs and the OAD Covariance Function with Automated Endmember Extraction 基于GPs和OAD协方差函数的高光谱图像分类与自动端元提取
Pub Date : 2011-11-07 DOI: 10.1109/ICTAI.2011.189
S. Schneider, A. Melkumyan, R. Murphy, E. Nettleton
In this paper we use a machine learning algorithm based on Gaussian Processes (GPs) and the Observation Angle Dependent (OAD) covariance function to classify hyper spectral imagery for the first time. This paper demonstrates the potential of the GP-OAD method for use in autonomous mining to identify and map geology and mineralogy on a vertical mine face. We discuss the importance of independent training data (i.e. a spectral library) to map any mine face without a priori knowledge. We compare an independent spectral library to other libraries, based on image data, and evaluate their relative performances to distinguish ore bearing zones from waste. Results show that the algorithm yields high accuracies (90%) and F-scores (77%), the best results are achieved when libraries are combined. We also demonstrate mapping of geology using imagery under different conditions of illumination (e.g. shade).
本文首次采用基于高斯过程(GPs)和观测角度相关(OAD)协方差函数的机器学习算法对高光谱图像进行分类。本文展示了GP-OAD方法在自主采矿中识别和绘制垂直工作面地质和矿物学图的潜力。我们讨论了独立训练数据(即光谱库)在没有先验知识的情况下绘制任何矿面的重要性。我们将一个独立的光谱库与其他基于图像数据的光谱库进行比较,并评估它们在区分含矿带和废矿带方面的相对性能。结果表明,该算法具有较高的准确率(90%)和f分数(77%),其中组合库效果最好。我们还演示了在不同光照条件下(例如阴影)使用图像绘制地质图。
{"title":"Classification of Hyperspectral Imagery Using GPs and the OAD Covariance Function with Automated Endmember Extraction","authors":"S. Schneider, A. Melkumyan, R. Murphy, E. Nettleton","doi":"10.1109/ICTAI.2011.189","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.189","url":null,"abstract":"In this paper we use a machine learning algorithm based on Gaussian Processes (GPs) and the Observation Angle Dependent (OAD) covariance function to classify hyper spectral imagery for the first time. This paper demonstrates the potential of the GP-OAD method for use in autonomous mining to identify and map geology and mineralogy on a vertical mine face. We discuss the importance of independent training data (i.e. a spectral library) to map any mine face without a priori knowledge. We compare an independent spectral library to other libraries, based on image data, and evaluate their relative performances to distinguish ore bearing zones from waste. Results show that the algorithm yields high accuracies (90%) and F-scores (77%), the best results are achieved when libraries are combined. We also demonstrate mapping of geology using imagery under different conditions of illumination (e.g. shade).","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"105 8","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121014123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Using the H-Divergence to Prune Probabilistic Automata 利用h散度对概率自动机进行剪枝
Pub Date : 2011-11-07 DOI: 10.1109/ICTAI.2011.114
Marc Bernard, Baptiste Jeudy, Jean-Philippe Peyrache, M. Sebban, F. Thollard
A problem usually encountered in probabilistic automata learning is the difficulty to deal with large training samples and/or wide alphabets. This is partially due to the size of the resulting Probabilistic Prefix Tree (PPT) from which state merging-based learning algorithms are generally applied. In this paper, we propose a novel method to prune PPTs by making use of the H-divergence d_H, recently introduced in the field of domain adaptation. d_H is based on the classification error made by an hypothesis learned from unlabeled examples drawn according to two distributions to compare. Through a thorough comparison with state-of-the-art divergence measures, we provide experimental evidences that demonstrate the efficiency of our method based on this simple and intuitive criterion.
在概率自动机学习中经常遇到的一个问题是难以处理大型训练样本和/或广泛的字母。这部分是由于结果的概率前缀树(PPT)的大小,通常应用基于状态合并的学习算法。本文提出了一种利用域自适应领域新近引入的h -散度d_H对PPTs进行剪枝的新方法。d_H是基于从根据两个分布进行比较的未标记示例中学习到的假设所产生的分类误差。通过与最先进的散度测量方法的全面比较,我们提供了实验证据,证明了基于该简单直观准则的方法的有效性。
{"title":"Using the H-Divergence to Prune Probabilistic Automata","authors":"Marc Bernard, Baptiste Jeudy, Jean-Philippe Peyrache, M. Sebban, F. Thollard","doi":"10.1109/ICTAI.2011.114","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.114","url":null,"abstract":"A problem usually encountered in probabilistic automata learning is the difficulty to deal with large training samples and/or wide alphabets. This is partially due to the size of the resulting Probabilistic Prefix Tree (PPT) from which state merging-based learning algorithms are generally applied. In this paper, we propose a novel method to prune PPTs by making use of the H-divergence d_H, recently introduced in the field of domain adaptation. d_H is based on the classification error made by an hypothesis learned from unlabeled examples drawn according to two distributions to compare. Through a thorough comparison with state-of-the-art divergence measures, we provide experimental evidences that demonstrate the efficiency of our method based on this simple and intuitive criterion.","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128208730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ReadAid: A Robust and Fully-Automated Readability Assessment Tool ReadAid:一个健壮的全自动可读性评估工具
Rani Qumsiyeh, Yiu-Kai Ng
Reading is an integral part of educational development, however, it is frustrating for people who struggle to understand (are not motivated to read, respectively) text documents that are beyond (below, respectively) their readability levels. Finding appropriate reading materials, with or without first scanning through their contents, is a challenge, since there are tremendous amount of documents these days and a clear majority of them are not tagged with their readability levels. Even though existing readability assessment tools determine readability levels of text documents, they analyze solely the lexical, syntactic, and/or semantic properties of a document, which are neither fully-automated, generalized, nor well-defined and are mostly based on observations. To advance the current readability analysis technique, we propose a robust, fully-automated readability analyzer, denoted ReadAid, which employs support vector machines to combine features from the US Curriculum and College Board, traditional readability measures, and the author(s) and subject area(s) of a text document d to assess the readability level of d. ReadAid can be applied for (i) filtering documents (retrieved in response to a web query) of a particular readability level, (ii) determining the readability levels of digitalized text documents, such as book chapters, magazine articles, and news stories, or (iii) dynamically analyzing, in real time, the grade level of a text document being created. The novelty of ReadAid lies on using authorship, subject areas, and academic concepts and grammatical constructions extracted from the US Curriculum to determine the readability level of a text document. Experimental results show that ReadAid is highly effective and outperforms existing state-of-the-art readability assessment tools.
阅读是教育发展的一个组成部分,然而,对于那些努力理解(没有动力阅读,分别)超出(低于,分别)可读性水平的文本文档的人来说,这是令人沮丧的。找到合适的阅读材料,不管是否先浏览一下它们的内容,都是一个挑战,因为现在有大量的文档,而且很明显大多数文档都没有标记它们的可读性级别。尽管现有的可读性评估工具确定了文本文档的可读性级别,但它们仅分析文档的词法、语法和/或语义属性,这些属性既不是全自动的、一般化的,也不是定义良好的,而且主要基于观察。为了推进当前的可读性分析技术,我们提出了一个鲁棒的、全自动的可读性分析器,称为ReadAid,它使用支持向量机将来自美国课程和大学理事会的特征、传统的可读性度量以及文本文档d的作者和主题领域结合起来,以评估d的可读性水平。ReadAid可以应用于(i)过滤特定可读性水平的文档(响应web查询检索);(ii)确定数字化文本文档的可读性水平,如书籍章节、杂志文章和新闻故事,或(iii)实时动态分析正在创建的文本文档的等级水平。ReadAid的新颖之处在于使用作者身份、学科领域、从美国课程中提取的学术概念和语法结构来确定文本文档的可读性水平。实验结果表明,ReadAid是一种高效的可读性评估工具。
{"title":"ReadAid: A Robust and Fully-Automated Readability Assessment Tool","authors":"Rani Qumsiyeh, Yiu-Kai Ng","doi":"10.1109/ICTAI.2011.87","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.87","url":null,"abstract":"Reading is an integral part of educational development, however, it is frustrating for people who struggle to understand (are not motivated to read, respectively) text documents that are beyond (below, respectively) their readability levels. Finding appropriate reading materials, with or without first scanning through their contents, is a challenge, since there are tremendous amount of documents these days and a clear majority of them are not tagged with their readability levels. Even though existing readability assessment tools determine readability levels of text documents, they analyze solely the lexical, syntactic, and/or semantic properties of a document, which are neither fully-automated, generalized, nor well-defined and are mostly based on observations. To advance the current readability analysis technique, we propose a robust, fully-automated readability analyzer, denoted ReadAid, which employs support vector machines to combine features from the US Curriculum and College Board, traditional readability measures, and the author(s) and subject area(s) of a text document d to assess the readability level of d. ReadAid can be applied for (i) filtering documents (retrieved in response to a web query) of a particular readability level, (ii) determining the readability levels of digitalized text documents, such as book chapters, magazine articles, and news stories, or (iii) dynamically analyzing, in real time, the grade level of a text document being created. The novelty of ReadAid lies on using authorship, subject areas, and academic concepts and grammatical constructions extracted from the US Curriculum to determine the readability level of a text document. Experimental results show that ReadAid is highly effective and outperforms existing state-of-the-art readability assessment tools.","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128351711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
New Computational Aspects in Master-Slave Systems of Semantic Schemas 语义图式主从系统的新计算方法
Pub Date : 2011-11-07 DOI: 10.1109/ICTAI.2011.105
N. Tandareanu, Cristina Zamfir
In this paper we reconsider the computations accomplished in a semantic schema. We reconsider also the computations in a master-slave systems of semantic schemas introduced in [6] as a cooperating system of such structures. We show that a master-slave system is adequate to represent distributed knowledge. To relieve this fact we describe such a system named DiSys implemented in Java by client-server technology.
在本文中,我们重新考虑了在语义模式中完成的计算。我们还将[6]中引入的语义模式主从系统中的计算重新考虑为这种结构的协作系统。我们证明了主从系统足以表示分布式知识。为了减轻这一事实,我们描述了一个用Java实现的名为DiSys的系统,该系统采用客户机-服务器技术。
{"title":"New Computational Aspects in Master-Slave Systems of Semantic Schemas","authors":"N. Tandareanu, Cristina Zamfir","doi":"10.1109/ICTAI.2011.105","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.105","url":null,"abstract":"In this paper we reconsider the computations accomplished in a semantic schema. We reconsider also the computations in a master-slave systems of semantic schemas introduced in [6] as a cooperating system of such structures. We show that a master-slave system is adequate to represent distributed knowledge. To relieve this fact we describe such a system named DiSys implemented in Java by client-server technology.","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"575 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133321847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Design and Development of a Social Intelligence Based Context-Aware Middleware Using BlackBoard 基于社会智能的上下文感知中间件的设计与开发
Pub Date : 2011-11-07 DOI: 10.1109/ICTAI.2011.151
Joohee Suh, Chong-woo Woo
The context-aware computing environment is changing due to recent development of the new computing devices and new concept of services. The systems are developing rapidly, but most of them focus on recognition of the collected information, not on the intelligent capability. In this paper, we defined a new context aware computing environment based on the concept of social intelligence, which implies an ability to share or utilize information by making a relationship, recognizes context by making an inference, and works in collaboration to offer services more efficiently. We have designed and developed a Social Intelligence based Context-Aware Middleware (SI-CAM), under the environment. The SI-CAM provides a service with following functions, multi context-awareness, context based task planning, and grouping intelligent entities for collaboration. The system is developed with blackboard based structure, and tested on virtual environment in the domain of ubiquitous restaurant. The experiment showed some significant results.
由于新的计算设备和新的服务概念的发展,上下文感知计算环境正在发生变化。目前系统发展迅速,但大多侧重于对采集信息的识别,而不是智能能力。在本文中,我们基于社会智能的概念定义了一个新的上下文感知计算环境,这意味着通过建立关系来共享或利用信息的能力,通过推理来识别上下文,并通过协作来更有效地提供服务。在该环境下,我们设计并开发了一个基于社会智能的上下文感知中间件(SI-CAM)。SI-CAM提供的服务具有以下功能:多上下文感知、基于上下文的任务规划和为协作对智能实体进行分组。该系统采用基于黑板的结构进行开发,并在无所不在的餐厅领域的虚拟环境中进行了测试。实验显示了一些重要的结果。
{"title":"Design and Development of a Social Intelligence Based Context-Aware Middleware Using BlackBoard","authors":"Joohee Suh, Chong-woo Woo","doi":"10.1109/ICTAI.2011.151","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.151","url":null,"abstract":"The context-aware computing environment is changing due to recent development of the new computing devices and new concept of services. The systems are developing rapidly, but most of them focus on recognition of the collected information, not on the intelligent capability. In this paper, we defined a new context aware computing environment based on the concept of social intelligence, which implies an ability to share or utilize information by making a relationship, recognizes context by making an inference, and works in collaboration to offer services more efficiently. We have designed and developed a Social Intelligence based Context-Aware Middleware (SI-CAM), under the environment. The SI-CAM provides a service with following functions, multi context-awareness, context based task planning, and grouping intelligent entities for collaboration. The system is developed with blackboard based structure, and tested on virtual environment in the domain of ubiquitous restaurant. The experiment showed some significant results.","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128844407","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Decomposition-Based Approach to OWL DL Ontology Diagnosis 基于分解的OWL DL本体诊断方法
Pub Date : 2011-11-07 DOI: 10.1109/ICTAI.2011.104
Jianfeng Du, G. Qi, Jeff Z. Pan, Yi-Dong Shen
Computing all diagnoses of an inconsistent ontology is important in ontology-based applications. However, the number of diagnoses can be very large. It is impractical to enumerate all diagnoses before identifying the target one to render the ontology consistent. Hence, we propose to represent all diagnoses by multiple sets of partial diagnoses, where the total number of partial diagnoses can be small and the target diagnosis can be directly retrieved from these partial diagnoses. We also propose methods for computing the new representation of all diagnoses in an OWL DL ontology. Experimental results show that computing the new representation of all diagnoses is much easier than directly computing all diagnoses.
在基于本体的应用中,计算不一致本体的所有诊断是非常重要的。然而,诊断的数量可能非常大。在确定目标诊断之前列举所有诊断以使本体保持一致是不切实际的。因此,我们建议用多组部分诊断来表示所有诊断,其中部分诊断的总数可以很小,并且可以直接从这些部分诊断中检索到目标诊断。我们还提出了计算OWL DL本体中所有诊断的新表示的方法。实验结果表明,计算所有诊断的新表示比直接计算所有诊断要容易得多。
{"title":"A Decomposition-Based Approach to OWL DL Ontology Diagnosis","authors":"Jianfeng Du, G. Qi, Jeff Z. Pan, Yi-Dong Shen","doi":"10.1109/ICTAI.2011.104","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.104","url":null,"abstract":"Computing all diagnoses of an inconsistent ontology is important in ontology-based applications. However, the number of diagnoses can be very large. It is impractical to enumerate all diagnoses before identifying the target one to render the ontology consistent. Hence, we propose to represent all diagnoses by multiple sets of partial diagnoses, where the total number of partial diagnoses can be small and the target diagnosis can be directly retrieved from these partial diagnoses. We also propose methods for computing the new representation of all diagnoses in an OWL DL ontology. Experimental results show that computing the new representation of all diagnoses is much easier than directly computing all diagnoses.","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117148724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
How to Reason by HeaRT in a Semantic Knowledge-Based Wiki 如何在语义知识维基中进行心灵推理
W. T. Adrian, Szymon Bobek, G. J. Nalepa, K. Kaczor, Krzysztof Kluza
Semantic wikis constitute an increasingly popular class of systems for collaborative knowledge engineering. We developed Loki, a semantic wiki that uses a logic-based knowledge representation. It is compatible with semantic annotations mechanism as well as Semantic Web languages. We integrated the system with a rule engine called Heart that supports inference with production rules. Several modes for modularized rule bases, suitable for the distributed rule bases present in a wiki, are considered. Embedding the rule engine enables strong reasoning and allows to run production rules over semantic knowledge bases. In the paper, we demonstrate the system concepts and functionality using an illustrative example.
语义wiki构成了协作知识工程中日益流行的一类系统。我们开发了Loki,一个使用基于逻辑的知识表示的语义wiki。它兼容语义注释机制和语义Web语言。我们将该系统与一个名为Heart的规则引擎集成在一起,该引擎支持使用生产规则进行推理。考虑了适合wiki中分布式规则库的模块化规则库的几种模式。嵌入规则引擎支持强大的推理,并允许在语义知识库上运行生产规则。在本文中,我们用一个说明性的例子来说明系统的概念和功能。
{"title":"How to Reason by HeaRT in a Semantic Knowledge-Based Wiki","authors":"W. T. Adrian, Szymon Bobek, G. J. Nalepa, K. Kaczor, Krzysztof Kluza","doi":"10.1109/ICTAI.2011.71","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.71","url":null,"abstract":"Semantic wikis constitute an increasingly popular class of systems for collaborative knowledge engineering. We developed Loki, a semantic wiki that uses a logic-based knowledge representation. It is compatible with semantic annotations mechanism as well as Semantic Web languages. We integrated the system with a rule engine called Heart that supports inference with production rules. Several modes for modularized rule bases, suitable for the distributed rule bases present in a wiki, are considered. Embedding the rule engine enables strong reasoning and allows to run production rules over semantic knowledge bases. In the paper, we demonstrate the system concepts and functionality using an illustrative example.","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115264064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
The Effect of the Characteristics of the Dataset on the Selection Stability 数据集特征对选择稳定性的影响
Pub Date : 2011-11-07 DOI: 10.1109/ICTAI.2011.167
Salem Alelyani, Huan Liu, Lei Wang
Feature selection is an effective technique to reduce the dimensionality of a data set and to select relevant features for the domain problem. Recently, stability of feature selection methods has gained increasing attention. In fact, it has become a crucial factor in determining the goodness of a feature selection algorithm besides the learning performance. In this work, we conduct an extensive experimental study using verity of data sets and different well-known feature selection algorithms in order to study the behavior of these algorithms in terms of the stability.
特征选择是一种有效的降低数据集的维数并为领域问题选择相关特征的技术。近年来,特征选择方法的稳定性越来越受到人们的关注。事实上,除了学习性能之外,它已经成为决定特征选择算法好坏的关键因素。在这项工作中,我们使用数据集的真实性和不同的知名特征选择算法进行了广泛的实验研究,以研究这些算法在稳定性方面的行为。
{"title":"The Effect of the Characteristics of the Dataset on the Selection Stability","authors":"Salem Alelyani, Huan Liu, Lei Wang","doi":"10.1109/ICTAI.2011.167","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.167","url":null,"abstract":"Feature selection is an effective technique to reduce the dimensionality of a data set and to select relevant features for the domain problem. Recently, stability of feature selection methods has gained increasing attention. In fact, it has become a crucial factor in determining the goodness of a feature selection algorithm besides the learning performance. In this work, we conduct an extensive experimental study using verity of data sets and different well-known feature selection algorithms in order to study the behavior of these algorithms in terms of the stability.","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115281852","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Ranking in Co-effecting Multi-object/Link Types Networks 协同效应多对象/链路类型网络的排序
Bo Zhou, Manna Wu, Xin Xia, Chao Wu
Research on link based object ranking attracts increasing attention these years, which also brings computer science research and business marketing brand-new concepts, opportunities as well as a great deal of challenges. With prosperity of web pages search engine and widely use of social networks, recent graph-theoretic ranking approaches have achieved remarkable successes although most of them are focus on homogeneous networks studying. Previous study on co-ranking methods tries to divide heterogeneous networks into multiple homogeneous sub-networks and ties between different sub-networks. This paper proposes an efficient topic biased ranking method for bringing order to co-effecting heterogeneous networks among authors, papers and accepted institutions (journals/conferences) within one single random surfer. This new method aims to update ranks for different types of objects (author, paper, journals/conferences) at each random walk.
近年来,基于链接的对象排序研究越来越受到人们的关注,这给计算机科学研究和商业营销带来了全新的概念和机遇,同时也带来了许多挑战。随着网页搜索引擎的蓬勃发展和社交网络的广泛应用,近年来的图论排序方法虽然大多集中在同质网络的研究上,但也取得了显著的成功。以往的协同排序方法试图将异构网络划分为多个同质子网络,并将不同子网络之间的联系进行划分。本文提出了一种有效的主题偏向排序方法,用于在单个随机冲浪者中对作者、论文和被接受的机构(期刊/会议)之间的协同效应异构网络进行排序。这种新方法旨在在每次随机漫步时更新不同类型对象(作者、论文、期刊/会议)的排名。
{"title":"Ranking in Co-effecting Multi-object/Link Types Networks","authors":"Bo Zhou, Manna Wu, Xin Xia, Chao Wu","doi":"10.1109/ICTAI.2011.84","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.84","url":null,"abstract":"Research on link based object ranking attracts increasing attention these years, which also brings computer science research and business marketing brand-new concepts, opportunities as well as a great deal of challenges. With prosperity of web pages search engine and widely use of social networks, recent graph-theoretic ranking approaches have achieved remarkable successes although most of them are focus on homogeneous networks studying. Previous study on co-ranking methods tries to divide heterogeneous networks into multiple homogeneous sub-networks and ties between different sub-networks. This paper proposes an efficient topic biased ranking method for bringing order to co-effecting heterogeneous networks among authors, papers and accepted institutions (journals/conferences) within one single random surfer. This new method aims to update ranks for different types of objects (author, paper, journals/conferences) at each random walk.","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127341193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Graph-Based Bilingual Corpus Selection with Sentence Pair Ranking for Statistical Machine Translation 基于统计机器翻译的改进图双语语料库选择与句子对排序
Wen-Han Chao, Zhoujun Li
In statistical machine translation, the number of sentence pairs in the bilingual corpus is very important to the quality of translation. However, when the quantity reaches some extent, enlarging corpus has less effect on the translation, whereas increasing greatly the time and space complexity to building translation systems, which hinders the development of statistical machine translation. In this paper, we propose several ranking approaches to measure the quantity of information of each sentence pair, and apply them into a graph-based bilingual corpus selection framework to form an improved corpus selection approach, which now considers the difference of the initial quantities of information between the sentence pairs. Our experiments in a Chinese-English translation task show that, selecting only 50% of the whole corpus via the graph-based selection approach as training set, we can obtain the near translation result with the one using the whole corpus, and we obtain better results than the baselines after using the IDF-related ranking approach.
在统计机器翻译中,双语语料库中句子对的数量对翻译质量至关重要。然而,当数量达到一定程度时,扩大语料库对翻译的影响不大,反而大大增加了构建翻译系统的时间和空间复杂性,阻碍了统计机器翻译的发展。本文提出了几种衡量句子对信息量的排序方法,并将其应用到基于图的双语语料库选择框架中,形成了一种考虑句子对初始信息量差异的改进语料库选择方法。我们在汉英翻译任务中的实验表明,通过基于图的选择方法只选择整个语料库的50%作为训练集,我们可以获得与使用整个语料库的翻译结果接近的翻译结果,并且使用idf相关排序方法获得比基线更好的结果。
{"title":"Improved Graph-Based Bilingual Corpus Selection with Sentence Pair Ranking for Statistical Machine Translation","authors":"Wen-Han Chao, Zhoujun Li","doi":"10.1109/ICTAI.2011.73","DOIUrl":"https://doi.org/10.1109/ICTAI.2011.73","url":null,"abstract":"In statistical machine translation, the number of sentence pairs in the bilingual corpus is very important to the quality of translation. However, when the quantity reaches some extent, enlarging corpus has less effect on the translation, whereas increasing greatly the time and space complexity to building translation systems, which hinders the development of statistical machine translation. In this paper, we propose several ranking approaches to measure the quantity of information of each sentence pair, and apply them into a graph-based bilingual corpus selection framework to form an improved corpus selection approach, which now considers the difference of the initial quantities of information between the sentence pairs. Our experiments in a Chinese-English translation task show that, selecting only 50% of the whole corpus via the graph-based selection approach as training set, we can obtain the near translation result with the one using the whole corpus, and we obtain better results than the baselines after using the IDF-related ranking approach.","PeriodicalId":332661,"journal":{"name":"2011 IEEE 23rd International Conference on Tools with Artificial Intelligence","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125992505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2011 IEEE 23rd International Conference on Tools with Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1