首页 > 最新文献

2008 IEEE International Conference on Information Reuse and Integration最新文献

英文 中文
Optimizing lineage information in genetic algorithms for producing superior models 优化遗传算法中的谱系信息以产生优越的模型
Pub Date : 2008-07-13 DOI: 10.1109/IRI.2008.4583049
G. Boetticher, J. Rudisill
A lot of research in the area of genetic algorithms (GA) is applied, but little research examines the impact of lineage information in optimizing a GA. Normally, researchers consider primarily elitism, an approach which carries only a very small fixed subset of the population to the next generation, as a lineage strategy. This paper investigates several different lineage percentages (what percent of the population to carry forward) to determine an ideal percentage or range from improving the accuracy of a GA. Several experiments are performed, and all results are statistically validated.
在遗传算法领域有大量的研究,但很少有研究考察谱系信息对遗传算法优化的影响。通常,研究人员主要认为精英主义是一种谱系策略,这种方法只将人口中非常小的固定子集传递给下一代。本文研究了几种不同的谱系百分比(要继承的人口百分比),以确定一个理想的百分比或范围,以提高遗传算法的准确性。进行了多次实验,所有结果都得到了统计验证。
{"title":"Optimizing lineage information in genetic algorithms for producing superior models","authors":"G. Boetticher, J. Rudisill","doi":"10.1109/IRI.2008.4583049","DOIUrl":"https://doi.org/10.1109/IRI.2008.4583049","url":null,"abstract":"A lot of research in the area of genetic algorithms (GA) is applied, but little research examines the impact of lineage information in optimizing a GA. Normally, researchers consider primarily elitism, an approach which carries only a very small fixed subset of the population to the next generation, as a lineage strategy. This paper investigates several different lineage percentages (what percent of the population to carry forward) to determine an ideal percentage or range from improving the accuracy of a GA. Several experiments are performed, and all results are statistically validated.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"269 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125827651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Segmentation of medical images by region growing 基于区域增长的医学图像分割方法
Pub Date : 2008-07-13 DOI: 10.1109/IRI.2008.4583071
Itzel Abundez Barrera, Citlalih Gutierrez Estrada, S. D. Zagal, M. N. Perez
The possibility of combining different technologies and developing better performing systems that offer quality results leads to the creation of systems with enhanced adaptation and analysis. Such analysis allows the interaction with previously evaluated techniques, providing reliable, successful results. Such is the case for the research work detailed in this article, which is focused on a technique that allows reusing previously analyzed and formalized information with the support of the Unified Modeled Language (UML). Information is handled in modules aimed to generate the segmentation of medical images without intervention of a specialist. The purpose is to deliver regions of interest for the early detection of cervical cancer.
结合不同的技术和开发提供高质量结果的性能更好的系统的可能性导致创建具有增强的适应性和分析的系统。这种分析允许与先前评估的技术相互作用,提供可靠、成功的结果。这就是本文中详细介绍的研究工作的情况,它关注的是一种技术,该技术允许在统一建模语言(UML)的支持下重用以前分析过的和形式化的信息。信息在模块中处理,目的是在没有专家干预的情况下生成医学图像的分割。目的是为宫颈癌的早期发现提供感兴趣的区域。
{"title":"Segmentation of medical images by region growing","authors":"Itzel Abundez Barrera, Citlalih Gutierrez Estrada, S. D. Zagal, M. N. Perez","doi":"10.1109/IRI.2008.4583071","DOIUrl":"https://doi.org/10.1109/IRI.2008.4583071","url":null,"abstract":"The possibility of combining different technologies and developing better performing systems that offer quality results leads to the creation of systems with enhanced adaptation and analysis. Such analysis allows the interaction with previously evaluated techniques, providing reliable, successful results. Such is the case for the research work detailed in this article, which is focused on a technique that allows reusing previously analyzed and formalized information with the support of the Unified Modeled Language (UML). Information is handled in modules aimed to generate the segmentation of medical images without intervention of a specialist. The purpose is to deliver regions of interest for the early detection of cervical cancer.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125503330","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A conflict-based confidence measure for associative classification 关联分类中基于冲突的置信度度量
Pub Date : 2008-07-13 DOI: 10.1109/IRI.2008.4583039
P. Vateekul, M. Shyu
Associative classification has aroused significant attention recently and achieved promising results. In the rule ranking process, the confidence measure is usually used to sort the class association rules (CARs). However, it may be not good enough for a classification task due to a low discrimination power to instances in the other classes. In this paper, we propose a novel conflict-based confidence measure with an interleaving ranking strategy for re-ranking CARs in an associative classification framework, which better captures the conflict between a rule and a training data instance. In the experiments, the traditional confidence measure and our proposed conflict-based confidence measure with the interleaving ranking strategy are applied as the primary sorting criterion for CARs. The experimental results show that the proposed associative classification framework achieves promising classification accuracy with the use of the conflict-based confidence measure, particularly for an imbalanced data set.
联想分类近年来引起了广泛的关注,并取得了可喜的成果。在规则排序过程中,通常使用置信度对类关联规则(car)进行排序。然而,对于分类任务来说,它可能不够好,因为它对其他类中的实例的辨别能力很低。在本文中,我们提出了一种新的基于冲突的置信度度量,采用交错排序策略对关联分类框架中的car进行重新排序,从而更好地捕获规则与训练数据实例之间的冲突。在实验中,将传统的置信度度量和我们提出的基于冲突的置信度度量结合交错排序策略作为car的主要排序标准。实验结果表明,使用基于冲突的置信度度量,提出的关联分类框架获得了很好的分类精度,特别是对于不平衡数据集。
{"title":"A conflict-based confidence measure for associative classification","authors":"P. Vateekul, M. Shyu","doi":"10.1109/IRI.2008.4583039","DOIUrl":"https://doi.org/10.1109/IRI.2008.4583039","url":null,"abstract":"Associative classification has aroused significant attention recently and achieved promising results. In the rule ranking process, the confidence measure is usually used to sort the class association rules (CARs). However, it may be not good enough for a classification task due to a low discrimination power to instances in the other classes. In this paper, we propose a novel conflict-based confidence measure with an interleaving ranking strategy for re-ranking CARs in an associative classification framework, which better captures the conflict between a rule and a training data instance. In the experiments, the traditional confidence measure and our proposed conflict-based confidence measure with the interleaving ranking strategy are applied as the primary sorting criterion for CARs. The experimental results show that the proposed associative classification framework achieves promising classification accuracy with the use of the conflict-based confidence measure, particularly for an imbalanced data set.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122168574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A generic notification system for Internet information 一个通用的互联网信息通知系统
Pub Date : 2008-07-13 DOI: 10.1109/IRI.2008.4583024
N. Mohamed, J. Al-Jaroodi, I. Jawhar
The Internet provides a huge amount of online and dynamic information related to stock information, currency exchange rates, interest rates, expected weather status, oil prices, and many other topics. This information is publicly available on dynamic HTML documents or on web services. This paper discusses a flexible notification system that utilizes the available online information and allows users to define a set of notifications that they are interested in. Based on the defined notification conditions, users will be notified by email and/or SMS messages whenever one or more of the conditions are met. In this system, users use Java-based configurations to define the notification requirements. This system also solves some of the issues facing utilizing public information available on the Internet to build the needed notifications. This includes the problem of capturing highly dynamic Internet information as well as supporting advanced types of notifications.
Internet提供了大量与股票信息、货币汇率、利率、预期天气状况、石油价格和许多其他主题相关的在线和动态信息。这些信息可以在动态HTML文档或web服务上公开获得。本文讨论了一个灵活的通知系统,它利用可用的在线信息,并允许用户定义一组他们感兴趣的通知。根据定义的通知条件,当满足一个或多个条件时,将通过电子邮件和/或短信通知用户。在本系统中,用户使用基于java的配置来定义通知需求。该系统还解决了利用Internet上的公共信息构建所需通知所面临的一些问题。这包括捕获高度动态的Internet信息以及支持高级类型的通知的问题。
{"title":"A generic notification system for Internet information","authors":"N. Mohamed, J. Al-Jaroodi, I. Jawhar","doi":"10.1109/IRI.2008.4583024","DOIUrl":"https://doi.org/10.1109/IRI.2008.4583024","url":null,"abstract":"The Internet provides a huge amount of online and dynamic information related to stock information, currency exchange rates, interest rates, expected weather status, oil prices, and many other topics. This information is publicly available on dynamic HTML documents or on web services. This paper discusses a flexible notification system that utilizes the available online information and allows users to define a set of notifications that they are interested in. Based on the defined notification conditions, users will be notified by email and/or SMS messages whenever one or more of the conditions are met. In this system, users use Java-based configurations to define the notification requirements. This system also solves some of the issues facing utilizing public information available on the Internet to build the needed notifications. This includes the problem of capturing highly dynamic Internet information as well as supporting advanced types of notifications.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128263403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Hybrid sampling for imbalanced data 不平衡数据的混合采样
Pub Date : 2008-07-13 DOI: 10.3233/ICA-2009-0314
Chris Seiffert, T. Khoshgoftaar, J. V. Hulse
Decision tree learning in the presence of imbalanced data is an issue of great practical importance, as such data is ubiquitous in a wide variety of application domains. We propose hybrid data sampling, which uses a combination of two sampling techniques such as random oversampling and random undersampling, to create a balanced dataset for use in the construction of decision tree classification models. The results demonstrate that our methodology is often able to improve the performance of a C4.5 decision tree learner in the context of imbalanced data.
存在不平衡数据的决策树学习是一个具有重要实际意义的问题,因为这种数据在各种应用领域中无处不在。我们提出了混合数据采样,它结合了随机过采样和随机欠采样两种采样技术,以创建一个平衡的数据集,用于构建决策树分类模型。结果表明,我们的方法通常能够提高C4.5决策树学习器在不平衡数据背景下的性能。
{"title":"Hybrid sampling for imbalanced data","authors":"Chris Seiffert, T. Khoshgoftaar, J. V. Hulse","doi":"10.3233/ICA-2009-0314","DOIUrl":"https://doi.org/10.3233/ICA-2009-0314","url":null,"abstract":"Decision tree learning in the presence of imbalanced data is an issue of great practical importance, as such data is ubiquitous in a wide variety of application domains. We propose hybrid data sampling, which uses a combination of two sampling techniques such as random oversampling and random undersampling, to create a balanced dataset for use in the construction of decision tree classification models. The results demonstrate that our methodology is often able to improve the performance of a C4.5 decision tree learner in the context of imbalanced data.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130779641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 74
Rule randomization for propositional logic-based workflow verification 基于命题逻辑的工作流验证规则随机化
Pub Date : 2008-07-13 DOI: 10.1109/IRI.2008.4583060
Q. Liang, S. Rubin
Workflow verification has been a well studied research topic during the past few years. Theorem proof based approaches to workflow verification become popular due to several advantages including being based on formal characterization with rigorous and non-ambiguous inference mechanisms. However, a common problem to these inference mechanisms is combinatorial explosions, which forms a major performance hurdle to workflow verification systems based on inference. In this paper, we study how randomization enables reuse and reduces processing time in logic based workflow verification approaches. We, in particular, look at a propositional logic based workflow verification technique. For the logic inference rules, which are used to infer new truthful propositions from existing truthful propositions in this logic, we apply randomization to the inference rules after each verification task such that new inference rules reflecting the componentized verification are added to the inference rule sets. We reviewed the savings incurred in verifying a workflow pattern and provide a theoretical analysis.
在过去的几年中,工作流验证一直是一个研究得很好的研究课题。基于定理证明的工作流验证方法之所以流行,是因为它有几个优点,包括基于具有严格和非模糊推理机制的形式化表征。然而,这些推理机制的一个共同问题是组合爆炸,这构成了基于推理的工作流验证系统的主要性能障碍。在本文中,我们研究了随机化如何在基于逻辑的工作流验证方法中实现重用和减少处理时间。我们特别关注基于命题逻辑的工作流验证技术。对于逻辑推理规则,用于从该逻辑中已有的真实命题中推断出新的真实命题,我们在每个验证任务后对推理规则进行随机化处理,从而将反映组件化验证的新推理规则添加到推理规则集中。我们回顾了验证工作流模式所带来的节省,并提供了理论分析。
{"title":"Rule randomization for propositional logic-based workflow verification","authors":"Q. Liang, S. Rubin","doi":"10.1109/IRI.2008.4583060","DOIUrl":"https://doi.org/10.1109/IRI.2008.4583060","url":null,"abstract":"Workflow verification has been a well studied research topic during the past few years. Theorem proof based approaches to workflow verification become popular due to several advantages including being based on formal characterization with rigorous and non-ambiguous inference mechanisms. However, a common problem to these inference mechanisms is combinatorial explosions, which forms a major performance hurdle to workflow verification systems based on inference. In this paper, we study how randomization enables reuse and reduces processing time in logic based workflow verification approaches. We, in particular, look at a propositional logic based workflow verification technique. For the logic inference rules, which are used to infer new truthful propositions from existing truthful propositions in this logic, we apply randomization to the inference rules after each verification task such that new inference rules reflecting the componentized verification are added to the inference rule sets. We reviewed the savings incurred in verifying a workflow pattern and provide a theoretical analysis.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"97 38","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131879417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A word-based predictive text entry method for Khmer language 高棉语基于词的预测文本输入方法
Pub Date : 2008-07-13 DOI: 10.1109/IRI.2008.4583032
Phavy Ouk, Ye Kyaw Thu, M. Matsumoto, Y. Urano
This paper begins with a discussion on the difficulties in applying word-based text entry method for Khmer, official language of Cambodia. Then, we propose a word-based predictive method based on careful analysis on the structure of current Khmer typing system. To evaluate the proposed text entry, we designed and implemented two interface prototypes; the 12-key mobile phone interface and the stylus-based device interface such as Personal Digital Assistant (PDA). Results show that compared to the existing methods our model provides better keystrokes and speed.
本文首先讨论了在柬埔寨官方语言高棉语中应用基于词的文本录入方法的困难。然后,在分析当前高棉语打字系统结构的基础上,提出了一种基于词的预测方法。为了评估建议的文本输入,我们设计并实现了两个界面原型;12键手机界面和PDA (Personal Digital Assistant)等触控笔设备界面。结果表明,与现有的方法相比,我们的模型提供了更好的击键和速度。
{"title":"A word-based predictive text entry method for Khmer language","authors":"Phavy Ouk, Ye Kyaw Thu, M. Matsumoto, Y. Urano","doi":"10.1109/IRI.2008.4583032","DOIUrl":"https://doi.org/10.1109/IRI.2008.4583032","url":null,"abstract":"This paper begins with a discussion on the difficulties in applying word-based text entry method for Khmer, official language of Cambodia. Then, we propose a word-based predictive method based on careful analysis on the structure of current Khmer typing system. To evaluate the proposed text entry, we designed and implemented two interface prototypes; the 12-key mobile phone interface and the stylus-based device interface such as Personal Digital Assistant (PDA). Results show that compared to the existing methods our model provides better keystrokes and speed.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116311179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An entropy-based method for assessing the number of spatial outliers 一种基于熵的空间异常值数量评估方法
Pub Date : 2008-07-13 DOI: 10.1109/IRI.2008.4583037
Xutong Liu, Chang-Tien Lu, F. Chen
A spatial outlier is a spatial object whose non-spatial attributes are significantly different from those of its spatial neighbors. A major limitation associated with the existing outlier detection algorithms is that they generally require a pre-specified number of spatial outliers. Estimating an appropriate number of outliers for a spatial data set is one of the critical issues for outlier analysis. This paper proposes an entropy-based method to address this problem. We define the function of spatial local contrast entropy. Based on the local contrast and local contrast probability that derived from non-spatial and spatial attributes, the spatial local contrast entropy can be computed. By incrementally removing outliers, the entropy value will keep decreasing until it becomes stable at a certain point, where an optimal number of outliers can be estimated. We considered both the single attribute and the multiple attributes of spatial objects. Experiments conducted on the US Housing data validated the effectiveness of our proposed approach.
空间离群点是指其非空间属性与其空间邻居显著不同的空间对象。现有离群点检测算法的一个主要限制是,它们通常需要预先指定数量的空间离群点。离群值分析的关键问题之一是对空间数据集估计适当数量的离群值。本文提出了一种基于熵的方法来解决这个问题。我们定义了空间局部对比熵的函数。基于非空间属性和空间属性得到的局部对比度和局部对比度概率,计算空间局部对比度熵。通过逐步去除离群值,熵值将不断减小,直到在某一点稳定下来,此时可以估计出最优的离群值数量。我们考虑了空间对象的单属性和多属性。对美国住房数据进行的实验验证了我们提出的方法的有效性。
{"title":"An entropy-based method for assessing the number of spatial outliers","authors":"Xutong Liu, Chang-Tien Lu, F. Chen","doi":"10.1109/IRI.2008.4583037","DOIUrl":"https://doi.org/10.1109/IRI.2008.4583037","url":null,"abstract":"A spatial outlier is a spatial object whose non-spatial attributes are significantly different from those of its spatial neighbors. A major limitation associated with the existing outlier detection algorithms is that they generally require a pre-specified number of spatial outliers. Estimating an appropriate number of outliers for a spatial data set is one of the critical issues for outlier analysis. This paper proposes an entropy-based method to address this problem. We define the function of spatial local contrast entropy. Based on the local contrast and local contrast probability that derived from non-spatial and spatial attributes, the spatial local contrast entropy can be computed. By incrementally removing outliers, the entropy value will keep decreasing until it becomes stable at a certain point, where an optimal number of outliers can be estimated. We considered both the single attribute and the multiple attributes of spatial objects. Experiments conducted on the US Housing data validated the effectiveness of our proposed approach.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122602556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Automatic product feature extraction from online product reviews using maximum entropy with lexical and syntactic features 基于词法和句法特征的最大熵在线产品评论特征自动提取
Pub Date : 2008-07-13 DOI: 10.1109/IRI.2008.4583038
G. Somprasertsri, P. Lalitrojwong
The task of product feature extraction is to find product features that customers refer to their topic reviews. It would be useful to characterize the opinions about the products. We propose an approach for product feature extraction by combining lexical and syntactic features with a maximum entropy model. For the underlying principle of maximum entropy, it prefers the uniform distributions if there is no external knowledge. Using a maximum entropy approach, firstly we extract the learning features from the annotated corpus, secondly we train the maximum entropy model, thirdly we use trained model to extract product features, and finally we apply a natural language processing technique in postprocessing step to discover the remaining product features. Our experimental results show that this approach is suitable for automatic product feature extraction.
产品特征提取的任务是找到客户在主题评论中引用的产品特征。描述一下对产品的看法是很有用的。我们提出了一种将词法和句法特征与最大熵模型相结合的产品特征提取方法。对于最大熵的基本原理,如果没有外部知识,它倾向于均匀分布。首先利用最大熵方法从标注的语料库中提取学习特征,然后训练最大熵模型,然后利用训练好的模型提取产品特征,最后在后处理步骤中应用自然语言处理技术发现剩余的产品特征。实验结果表明,该方法适用于产品特征的自动提取。
{"title":"Automatic product feature extraction from online product reviews using maximum entropy with lexical and syntactic features","authors":"G. Somprasertsri, P. Lalitrojwong","doi":"10.1109/IRI.2008.4583038","DOIUrl":"https://doi.org/10.1109/IRI.2008.4583038","url":null,"abstract":"The task of product feature extraction is to find product features that customers refer to their topic reviews. It would be useful to characterize the opinions about the products. We propose an approach for product feature extraction by combining lexical and syntactic features with a maximum entropy model. For the underlying principle of maximum entropy, it prefers the uniform distributions if there is no external knowledge. Using a maximum entropy approach, firstly we extract the learning features from the annotated corpus, secondly we train the maximum entropy model, thirdly we use trained model to extract product features, and finally we apply a natural language processing technique in postprocessing step to discover the remaining product features. Our experimental results show that this approach is suitable for automatic product feature extraction.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"111 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115179782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 44
DEVS model composition by system entity structure 按系统实体结构组成DEVS模型
Pub Date : 2008-07-13 DOI: 10.1109/IRI.2008.4583078
S. Cheon, Doohwan Kim, B. Zeigler
The structural knowledge of a system represented in system entity structure (SES) supports the handling of organization issue in model compositions. A pruning operation results in a reduced structure, pruned entity structure (PES) that the SES is pruned to meet design objectives. The PES is eventually synthesized into a simulation model by combining it with models in the model base. SES is implemented in XML Metadata using Java language and Sun’s Document Object Models (DOM) specification. The SESBuilder1 software supports natural language input for an SES definition and XML instance generation of its PES. The coupling information expressed in natural language with restricted syntax processed by SESBuilder is used to compose models. As a real example, the compositions of Discrete Event Specification (DEVS) generator models, representing the US Climate Normals, are presented. The example presents the natural language input and XML instances for PES including coupling information in the SESBuilder.
以系统实体结构(SES)表示的系统的结构知识支持处理模型组合中的组织问题。剪枝操作会产生一个简化的结构,即剪枝实体结构(PES), SES被剪枝以满足设计目标。将PES与模型库中的模型结合,最终合成为仿真模型。SES是使用Java语言和Sun的文档对象模型(DOM)规范在XML元数据中实现的。SESBuilder1软件支持SES定义的自然语言输入和PES的XML实例生成。利用SESBuilder处理的受限制语法的自然语言表达的耦合信息组成模型。作为一个实际的例子,给出了代表美国气候常态的离散事件规范(DEVS)发电机模型的组成。该示例展示了PES的自然语言输入和XML实例,包括SESBuilder中的耦合信息。
{"title":"DEVS model composition by system entity structure","authors":"S. Cheon, Doohwan Kim, B. Zeigler","doi":"10.1109/IRI.2008.4583078","DOIUrl":"https://doi.org/10.1109/IRI.2008.4583078","url":null,"abstract":"The structural knowledge of a system represented in system entity structure (SES) supports the handling of organization issue in model compositions. A pruning operation results in a reduced structure, pruned entity structure (PES) that the SES is pruned to meet design objectives. The PES is eventually synthesized into a simulation model by combining it with models in the model base. SES is implemented in XML Metadata using Java language and Sun’s Document Object Models (DOM) specification. The SESBuilder1 software supports natural language input for an SES definition and XML instance generation of its PES. The coupling information expressed in natural language with restricted syntax processed by SESBuilder is used to compose models. As a real example, the compositions of Discrete Event Specification (DEVS) generator models, representing the US Climate Normals, are presented. The example presents the natural language input and XML instances for PES including coupling information in the SESBuilder.","PeriodicalId":169554,"journal":{"name":"2008 IEEE International Conference on Information Reuse and Integration","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130128854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2008 IEEE International Conference on Information Reuse and Integration
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1