首页 > 最新文献

2014 IEEE International Conference on Semantic Computing最新文献

英文 中文
A Semantic Approach to Enterprise Information Integration 企业信息集成的语义方法
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.23
A. Katasonov, A. Lattunen
The emerging Internet of Things technologies enable enterprises to collect a variety of real-time data from the physical world, making a case for accessing, combining, interpreting, and distributing such data in real-time too. Enterprise Information Integration (EII) aims at providing tools for integrating data from multiple sources without having to first load all the data into a central warehouse, and, in so, for accessing live data. In this paper, we introduce a practical semantic EII solution, which, in addition to addressing the data virtualization and federation problems of EII, also provides data abstraction and data pipeline capabilities. This solution has been implemented as a software platform as well as applied in an operational enterprise system in the parking domain.
新兴的物联网技术使企业能够从物理世界中收集各种实时数据,并为实时访问、组合、解释和分发这些数据提供了理由。企业信息集成(Enterprise Information Integration, EII)旨在提供工具,用于集成来自多个数据源的数据,而无需首先将所有数据加载到中央仓库中,从而访问实时数据。在本文中,我们介绍了一个实用的语义EII解决方案,该解决方案除了解决EII的数据虚拟化和联邦问题外,还提供了数据抽象和数据管道功能。该解决方案已作为软件平台实现,并应用于停车场领域的运营企业系统中。
{"title":"A Semantic Approach to Enterprise Information Integration","authors":"A. Katasonov, A. Lattunen","doi":"10.1109/ICSC.2014.23","DOIUrl":"https://doi.org/10.1109/ICSC.2014.23","url":null,"abstract":"The emerging Internet of Things technologies enable enterprises to collect a variety of real-time data from the physical world, making a case for accessing, combining, interpreting, and distributing such data in real-time too. Enterprise Information Integration (EII) aims at providing tools for integrating data from multiple sources without having to first load all the data into a central warehouse, and, in so, for accessing live data. In this paper, we introduce a practical semantic EII solution, which, in addition to addressing the data virtualization and federation problems of EII, also provides data abstraction and data pipeline capabilities. This solution has been implemented as a software platform as well as applied in an operational enterprise system in the parking domain.","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123937724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Exploration and Analysis of Undocumented Processes Using Heterogeneous and Unstructured Business Data 使用异构和非结构化业务数据的未记录流程的探索和分析
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.24
Sebastian Pospiech, Sven Mielke, R. Mertens, K. Jagannath, Michael Städler
The business world has become more dynamic than ever before. Global competition and today's rapid pace of development in many fields has led to shorter time-to-market intervals, as well as more complex products and services. These developments do often imply impromptu changes to existing business processes. These dynamics are aggravated when unforeseen paths have to be taken like it is often the case when problems are solved in customer support situations. This leads to undocumented business processes which pose a serious problem for management. In order to cope with this problem the discipline of Process Mining has emerged. In classical Process Mining, event logs generated for example by workflow management systems are used to create a process model. In order for classical Process Mining to work, the process therefore has to be implemented in such a system, it just lacks documentation. The above mentioned impromptu changes and impromptu processes do, however, lack any such documentation. In many cases event logs do not exist, at least not in the strict sense of the definition. Instead, traces left by a process might include unstructured data, such as emails or notes in a human readable format. In this paper we will demonstrate how it is possible to search and locate processes that exist in a company, but that are neither documented, nor implemented in any business process management system. The idea is to use all data stores in a company to find a trace of a process instance and to reconstruct and visualize it. The trace of this single instance is then generalized to a process template that covers all instances of that process. This generalization step generates a description that can manually be adapted in order to fit all process instances. While retrieving instances from structured data can be described by simple queries, retrieving process steps from unstructured data often requires more elaborate approaches. Hence, we have modified a search-engine to combine a simple word-search with ad-hoc ontologies that allow for defining synonym relations on a query-by-query basis.
商业世界比以往任何时候都更有活力。全球竞争和当今许多领域的快速发展使得产品上市时间缩短,产品和服务也更加复杂。这些开发通常意味着对现有业务流程的临时更改。当必须采取不可预见的路径时,这些动态就会恶化,就像在客户支持情况下解决问题时经常出现的情况一样。这将导致未记录的业务流程,给管理带来严重的问题。为了解决这一问题,过程挖掘这一学科应运而生。在经典的流程挖掘中,例如由工作流管理系统生成的事件日志用于创建流程模型。为了使经典的流程挖掘工作,流程因此必须在这样的系统中实现,它只是缺乏文档。然而,上面提到的临时更改和临时流程确实缺乏任何此类文档。在许多情况下,事件日志并不存在,至少不是严格意义上的定义。相反,进程留下的痕迹可能包括非结构化数据,例如人类可读格式的电子邮件或笔记。在本文中,我们将演示如何搜索和定位存在于公司中,但在任何业务流程管理系统中既没有记录也没有实现的流程。其思想是使用公司中的所有数据存储来查找流程实例的跟踪,并对其进行重构和可视化。然后将该单个实例的跟踪推广到涵盖该流程所有实例的流程模板。这个泛化步骤生成了一个描述,可以手动调整该描述以适应所有流程实例。虽然从结构化数据中检索实例可以通过简单的查询来描述,但从非结构化数据中检索流程步骤通常需要更精细的方法。因此,我们修改了一个搜索引擎,将简单的单词搜索与特定的本体结合起来,这些本体允许在每个查询的基础上定义同义词关系。
{"title":"Exploration and Analysis of Undocumented Processes Using Heterogeneous and Unstructured Business Data","authors":"Sebastian Pospiech, Sven Mielke, R. Mertens, K. Jagannath, Michael Städler","doi":"10.1109/ICSC.2014.24","DOIUrl":"https://doi.org/10.1109/ICSC.2014.24","url":null,"abstract":"The business world has become more dynamic than ever before. Global competition and today's rapid pace of development in many fields has led to shorter time-to-market intervals, as well as more complex products and services. These developments do often imply impromptu changes to existing business processes. These dynamics are aggravated when unforeseen paths have to be taken like it is often the case when problems are solved in customer support situations. This leads to undocumented business processes which pose a serious problem for management. In order to cope with this problem the discipline of Process Mining has emerged. In classical Process Mining, event logs generated for example by workflow management systems are used to create a process model. In order for classical Process Mining to work, the process therefore has to be implemented in such a system, it just lacks documentation. The above mentioned impromptu changes and impromptu processes do, however, lack any such documentation. In many cases event logs do not exist, at least not in the strict sense of the definition. Instead, traces left by a process might include unstructured data, such as emails or notes in a human readable format. In this paper we will demonstrate how it is possible to search and locate processes that exist in a company, but that are neither documented, nor implemented in any business process management system. The idea is to use all data stores in a company to find a trace of a process instance and to reconstruct and visualize it. The trace of this single instance is then generalized to a process template that covers all instances of that process. This generalization step generates a description that can manually be adapted in order to fit all process instances. While retrieving instances from structured data can be described by simple queries, retrieving process steps from unstructured data often requires more elaborate approaches. Hence, we have modified a search-engine to combine a simple word-search with ad-hoc ontologies that allow for defining synonym relations on a query-by-query basis.","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"15 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131300084","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Building Distant Supervised Relation Extractors 构建远程监督关系提取器
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.15
Thiago Nunes, D. Schwabe
A well-known drawback in building machine learning semantic relation detectors for natural language is the lack of a large number of qualified training instances for the target relations in multiple languages. Even when good results are achieved, the datasets used by the state-of-the-art approaches are rarely published. In order to address these problems, this work presents an automatic approach to build multilingual semantic relation detectors through distant supervision combining two of the largest resources of structured and unstructured content available on the Web, DBpedia and Wikipedia. We map the DBpedia ontology back to the Wikipedia text to extract more than 100.000 training instances for more than 90 DBpedia relations for English and Portuguese languages without human intervention. First, we mine the Wikipedia articles to find candidate instances for relations described in the DBpedia ontology. Second, we preprocess and normalize the data filtering out irrelevant instances. Finally, we use the normalized data to construct regularized logistic regression detectors that achieve more than 80% of F-Measure for both English and Portuguese languages. In this paper, we also compare the impact of different types of features on the accuracy of the trained detector, demonstrating significant performance improvements when combining lexical, syntactic and semantic features. Both the datasets and the code used in this research are available online.
为自然语言构建机器学习语义关系检测器的一个众所周知的缺点是缺乏大量的多语言目标关系的合格训练实例。即使取得了良好的结果,最先进的方法所使用的数据集也很少发表。为了解决这些问题,这项工作提出了一种自动方法,通过远程监督来构建多语言语义关系检测器,该方法结合了Web上两个最大的结构化和非结构化内容资源,DBpedia和Wikipedia。我们将DBpedia本体映射回维基百科文本,在没有人为干预的情况下,为英语和葡萄牙语的90多个DBpedia关系提取超过100,000个训练实例。首先,我们挖掘维基百科文章,为DBpedia本体中描述的关系找到候选实例。其次,对数据进行预处理和规范化,过滤掉不相关的实例。最后,我们使用归一化数据构建正则化逻辑回归检测器,该检测器对英语和葡萄牙语都达到了80%以上的F-Measure。在本文中,我们还比较了不同类型的特征对训练检测器准确性的影响,表明当结合词汇、句法和语义特征时,性能有显著提高。本研究中使用的数据集和代码都可以在网上获得。
{"title":"Building Distant Supervised Relation Extractors","authors":"Thiago Nunes, D. Schwabe","doi":"10.1109/ICSC.2014.15","DOIUrl":"https://doi.org/10.1109/ICSC.2014.15","url":null,"abstract":"A well-known drawback in building machine learning semantic relation detectors for natural language is the lack of a large number of qualified training instances for the target relations in multiple languages. Even when good results are achieved, the datasets used by the state-of-the-art approaches are rarely published. In order to address these problems, this work presents an automatic approach to build multilingual semantic relation detectors through distant supervision combining two of the largest resources of structured and unstructured content available on the Web, DBpedia and Wikipedia. We map the DBpedia ontology back to the Wikipedia text to extract more than 100.000 training instances for more than 90 DBpedia relations for English and Portuguese languages without human intervention. First, we mine the Wikipedia articles to find candidate instances for relations described in the DBpedia ontology. Second, we preprocess and normalize the data filtering out irrelevant instances. Finally, we use the normalized data to construct regularized logistic regression detectors that achieve more than 80% of F-Measure for both English and Portuguese languages. In this paper, we also compare the impact of different types of features on the accuracy of the trained detector, demonstrating significant performance improvements when combining lexical, syntactic and semantic features. Both the datasets and the code used in this research are available online.","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116914540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Semantic End-to-End Process Constraint Modeling Framework 语义端到端流程约束建模框架
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.32
Shasha Liu, K. Kochut
Process constraint modeling and development, focusing on how to enforce the conformity of process constraints throughout its lifecycle, including design, deployment and runtime execution, remains a big challenge in the research area of model-driven development, especially when such constraints are considered in the composite Web services and workflow applications. By extending our previous work in process constraint ontology and process constraint language with the capability of exception definition and handling, we propose a semantic end-to-end process constraint modeling framework. In addition to constraint modeling at the design time, both static and dynamic verification in constraint's lifecycle are covered. While the former concentrates on syntactic, semantic and service specification verifications during design and deployment phases, the latter focuses on constraint verification during runtime, with the help of the underlying monitoring module. The scenarios of emergency reaction and nepotism in interview are considered and we illustrate how ontology and semantic reasoning are utilized in a constraint's whole lifecycle.
流程约束建模和开发关注于如何在整个生命周期(包括设计、部署和运行时执行)中强制执行流程约束的一致性,这在模型驱动开发的研究领域仍然是一个巨大的挑战,特别是当在组合Web服务和工作流应用程序中考虑这些约束时。通过扩展我们在过程约束本体和过程约束语言方面的工作,提供异常定义和处理能力,我们提出了一个语义端到端过程约束建模框架。除了设计时的约束建模之外,还涵盖了约束生命周期中的静态和动态验证。前者专注于设计和部署阶段的语法、语义和服务规范验证,后者则在底层监控模块的帮助下,专注于运行时期间的约束验证。考虑了面试中的紧急反应和裙带关系两种情景,并说明了在约束的整个生命周期中如何利用本体和语义推理。
{"title":"A Semantic End-to-End Process Constraint Modeling Framework","authors":"Shasha Liu, K. Kochut","doi":"10.1109/ICSC.2014.32","DOIUrl":"https://doi.org/10.1109/ICSC.2014.32","url":null,"abstract":"Process constraint modeling and development, focusing on how to enforce the conformity of process constraints throughout its lifecycle, including design, deployment and runtime execution, remains a big challenge in the research area of model-driven development, especially when such constraints are considered in the composite Web services and workflow applications. By extending our previous work in process constraint ontology and process constraint language with the capability of exception definition and handling, we propose a semantic end-to-end process constraint modeling framework. In addition to constraint modeling at the design time, both static and dynamic verification in constraint's lifecycle are covered. While the former concentrates on syntactic, semantic and service specification verifications during design and deployment phases, the latter focuses on constraint verification during runtime, with the help of the underlying monitoring module. The scenarios of emergency reaction and nepotism in interview are considered and we illustrate how ontology and semantic reasoning are utilized in a constraint's whole lifecycle.","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123729227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Anomaly Detection in Time Series Radiotherapy Treatment Data 放射治疗时间序列数据的异常检测
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.64
T. Sipes, H. Karimabadi, Steve B. Jiang, K. Moore, Nan Li, Joseph R. Barr
The work presented here resulted in a valuable innovative technology tool for automatic detection of catastrophic errors in cancer radiotherapy, adding an important safeguard for patient safety. We designed a tool for Dynamic Modeling and Prediction of Radiotherapy Treatment Deviations from Intended Plans (Smart Tool) to automatically detect and highlight potential errors in a radiotherapy treatment plan, based on the data from several thousand prostate cancer treatments at Moore Cancer Research Center at University of California San Diego. Smart Tool determines if the treatment parameters are valid, against a previously built Predictive Model of a Medical Error (PMME). Smart Tool has the following main features: 1) It communicates with a radiotherapy treatment management system, checking all the treatment parameters in the background prior to execution, and after the human expert QA is completed, 2) The anomalous treatment parameters, if any, are detected using an innovative intelligent algorithm in a completely automatic and unsupervised manner, 3) It is a self-learning and constantly evolving system, the model is dynamically updated with the new treatment data, 4) It incorporates expert knowledge through the feedback loop of the dynamic process which updates the model with any new false positives (FP) and false negatives (FN), 4) When an outlier treatment parameter is detected, Smart Tool works by preventing the plan execution and highlighting the parameter for human intervention, 5) It is aimed at catastrophic errors, not small errors.
本文的研究成果为癌症放射治疗中灾难性错误的自动检测提供了一种有价值的创新技术工具,为患者安全提供了重要保障。基于加州大学圣地亚哥分校摩尔癌症研究中心数千例前列腺癌治疗的数据,我们设计了一个用于放射治疗偏离预期计划的动态建模和预测工具(智能工具),以自动检测和突出放射治疗计划中的潜在错误。智能工具根据先前构建的医疗错误预测模型(PMME)确定治疗参数是否有效。智能工具具有以下主要特性:1)与放疗治疗管理系统通信,在执行前在后台检查所有治疗参数,并在人类专家QA完成后;2)使用创新的智能算法,以完全自动和无监督的方式检测异常治疗参数;3)它是一个自学习和不断进化的系统,模型随着新的治疗数据动态更新;4)它通过动态过程的反馈回路结合专家知识,该反馈回路用任何新的假阳性(FP)和假阴性(FN)更新模型,4)当检测到异常值处理参数时,智能工具通过阻止计划执行并突出显示人工干预参数来工作,5)它针对的是灾难性错误,而不是小错误。
{"title":"Anomaly Detection in Time Series Radiotherapy Treatment Data","authors":"T. Sipes, H. Karimabadi, Steve B. Jiang, K. Moore, Nan Li, Joseph R. Barr","doi":"10.1109/ICSC.2014.64","DOIUrl":"https://doi.org/10.1109/ICSC.2014.64","url":null,"abstract":"The work presented here resulted in a valuable innovative technology tool for automatic detection of catastrophic errors in cancer radiotherapy, adding an important safeguard for patient safety. We designed a tool for Dynamic Modeling and Prediction of Radiotherapy Treatment Deviations from Intended Plans (Smart Tool) to automatically detect and highlight potential errors in a radiotherapy treatment plan, based on the data from several thousand prostate cancer treatments at Moore Cancer Research Center at University of California San Diego. Smart Tool determines if the treatment parameters are valid, against a previously built Predictive Model of a Medical Error (PMME). Smart Tool has the following main features: 1) It communicates with a radiotherapy treatment management system, checking all the treatment parameters in the background prior to execution, and after the human expert QA is completed, 2) The anomalous treatment parameters, if any, are detected using an innovative intelligent algorithm in a completely automatic and unsupervised manner, 3) It is a self-learning and constantly evolving system, the model is dynamically updated with the new treatment data, 4) It incorporates expert knowledge through the feedback loop of the dynamic process which updates the model with any new false positives (FP) and false negatives (FN), 4) When an outlier treatment parameter is detected, Smart Tool works by preventing the plan execution and highlighting the parameter for human intervention, 5) It is aimed at catastrophic errors, not small errors.","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125500046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cloud Resource Auto-scaling System Based on Hidden Markov Model (HMM) 基于隐马尔可夫模型(HMM)的云资源自动伸缩系统
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.43
A. Nikravesh, S. Ajila, Chung-Horng Lung
The elasticity characteristic of cloud computing enables clients to acquire and release resources on demand. This characteristic reduces clients' cost by making them pay for the resources they actually have used. On the other hand, clients are obligated to maintain Service Level Agreement (SLA) with their users. One approach to deal with this cost-performance trade-off is employing an auto-scaling system which automatically adjusts application's resources based on its load. In this paper we have proposed an auto-scaling system based on Hidden Markov Model (HMM). We have conducted an experiment on Amazon EC2 infrastructure to evaluate our model. Our results show HMM can generate correct scaling actions in 97% of time. CPU utilization, throughput, and response time are being considered as performance metrics in our experiment.
云计算的弹性特性使客户能够按需获取和释放资源。这一特性通过让客户为他们实际使用的资源付费来降低客户成本。另一方面,客户端有义务与用户维护服务水平协议(SLA)。处理这种成本-性能权衡的一种方法是采用自动伸缩系统,该系统根据应用程序的负载自动调整应用程序的资源。本文提出了一种基于隐马尔可夫模型(HMM)的自动缩放系统。我们在Amazon EC2基础设施上进行了一个实验来评估我们的模型。我们的结果表明HMM可以在97%的时间内生成正确的缩放动作。在我们的实验中,CPU利用率、吞吐量和响应时间被视为性能指标。
{"title":"Cloud Resource Auto-scaling System Based on Hidden Markov Model (HMM)","authors":"A. Nikravesh, S. Ajila, Chung-Horng Lung","doi":"10.1109/ICSC.2014.43","DOIUrl":"https://doi.org/10.1109/ICSC.2014.43","url":null,"abstract":"The elasticity characteristic of cloud computing enables clients to acquire and release resources on demand. This characteristic reduces clients' cost by making them pay for the resources they actually have used. On the other hand, clients are obligated to maintain Service Level Agreement (SLA) with their users. One approach to deal with this cost-performance trade-off is employing an auto-scaling system which automatically adjusts application's resources based on its load. In this paper we have proposed an auto-scaling system based on Hidden Markov Model (HMM). We have conducted an experiment on Amazon EC2 infrastructure to evaluate our model. Our results show HMM can generate correct scaling actions in 97% of time. CPU utilization, throughput, and response time are being considered as performance metrics in our experiment.","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116753300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
A Review on Semantic Web and Recent Trends in Its Applications 语义网的研究进展及其应用趋势
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.57
Oguzhan Menemencioglu, I. M. Orak
Semantic web works on producing machine readable data. So semantic web aims to overcome the amount of data that is consisted. The most important tool to access the data which exist in web is the search engine. Traditional search engines are insufficient in the face of the amount of data that is consisted as a result of the existing pages on the web. Semantic search engines are extensions to traditional search engines and improved version. This paper summarizes semantic web, traditional and semantic search engine concepts and infrastructure. Also semantic search approaches and differences from traditional approach are detailed. A summary of the literature is provided by touching on the trends on this area. In this respect, type of applications and the areas worked for are considered. Based on the data for two different years, trend on these points are analyzed and impacts of changes are discussed. It shows that evaluation on the semantic web continues and new applications and areas are also emerging.
语义网致力于产生机器可读的数据。所以语义网的目标是克服数据量的问题。访问网络中存在的数据的最重要的工具是搜索引擎。面对由网络上现有页面组成的数据量,传统的搜索引擎是不够的。语义搜索引擎是传统搜索引擎的扩展和改进版本。本文综述了语义网、传统搜索引擎和语义搜索引擎的概念和基础结构。并详细介绍了语义搜索方法及其与传统方法的区别。通过触及这一领域的趋势,提供了文献摘要。在这方面,审议了申请的类型和工作的领域。根据两个不同年份的数据,分析了这些点的变化趋势,并讨论了变化的影响。这表明对语义网的评价仍在继续,新的应用和领域也在不断涌现。
{"title":"A Review on Semantic Web and Recent Trends in Its Applications","authors":"Oguzhan Menemencioglu, I. M. Orak","doi":"10.1109/ICSC.2014.57","DOIUrl":"https://doi.org/10.1109/ICSC.2014.57","url":null,"abstract":"Semantic web works on producing machine readable data. So semantic web aims to overcome the amount of data that is consisted. The most important tool to access the data which exist in web is the search engine. Traditional search engines are insufficient in the face of the amount of data that is consisted as a result of the existing pages on the web. Semantic search engines are extensions to traditional search engines and improved version. This paper summarizes semantic web, traditional and semantic search engine concepts and infrastructure. Also semantic search approaches and differences from traditional approach are detailed. A summary of the literature is provided by touching on the trends on this area. In this respect, type of applications and the areas worked for are considered. Based on the data for two different years, trend on these points are analyzed and impacts of changes are discussed. It shows that evaluation on the semantic web continues and new applications and areas are also emerging.","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117249263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Context Infusion in Semantic Link Networks to Detect Cyber-attacks: A Flow-Based Detection Approach 语义链接网络中的上下文注入检测网络攻击:一种基于流的检测方法
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.29
Ahmed Aleroud, George Karabatis
Detection of cyber-attacks is a major responsibility for network managers and security specialists. Most existing Network Intrusion Detection systems rely on inspecting individual packets, an increasingly resource consuming task in today's high speed networks due to the overhead associated with accessing packet content. An alternative approach is to detect attack patterns by investigating IP flows. Since analyzing raw data extracted from IP flows lacks the semantic information needed to discover attacks, a novel approach is introduced that utilizes contextual information to semantically reveal cyber-attacks from IP flows. Time, location, and other contextual information mined from network flow data is utilized to create semantic links among alerts raised in response to suspicious flows. The semantic links are identified through an inference process on probabilistic semantic link networks (SLNs). The resulting links are used at run-time to retrieve relevant suspicious activities that represent possible steps in multi-step attacks.
检测网络攻击是网络管理人员和安全专家的主要职责。大多数现有的网络入侵检测系统依赖于检查单个数据包,由于访问数据包内容的开销,在当今的高速网络中,这是一项日益消耗资源的任务。另一种方法是通过调查IP流来检测攻击模式。由于分析从IP流中提取的原始数据缺乏发现攻击所需的语义信息,因此引入了一种利用上下文信息从语义上揭示IP流中的网络攻击的新方法。从网络流数据中挖掘的时间、位置和其他上下文信息用于在响应可疑流时发出的警报之间创建语义链接。在概率语义链路网络(sln)上,通过推理过程识别语义链路。生成的链接在运行时用于检索代表多步骤攻击中可能步骤的相关可疑活动。
{"title":"Context Infusion in Semantic Link Networks to Detect Cyber-attacks: A Flow-Based Detection Approach","authors":"Ahmed Aleroud, George Karabatis","doi":"10.1109/ICSC.2014.29","DOIUrl":"https://doi.org/10.1109/ICSC.2014.29","url":null,"abstract":"Detection of cyber-attacks is a major responsibility for network managers and security specialists. Most existing Network Intrusion Detection systems rely on inspecting individual packets, an increasingly resource consuming task in today's high speed networks due to the overhead associated with accessing packet content. An alternative approach is to detect attack patterns by investigating IP flows. Since analyzing raw data extracted from IP flows lacks the semantic information needed to discover attacks, a novel approach is introduced that utilizes contextual information to semantically reveal cyber-attacks from IP flows. Time, location, and other contextual information mined from network flow data is utilized to create semantic links among alerts raised in response to suspicious flows. The semantic links are identified through an inference process on probabilistic semantic link networks (SLNs). The resulting links are used at run-time to retrieve relevant suspicious activities that represent possible steps in multi-step attacks.","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122936255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
QuIET: A Text Classification Technique Using Automatically Generated Span Queries QuIET:使用自动生成跨度查询的文本分类技术
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.18
Vassilis Polychronopoulos, N. Pendar, S. Jeffery
We propose a novel algorithm, QuIET, for binary classification of texts. The method automatically generates a set of span queries from a set of annotated documents and uses the query set to categorize unlabeled texts. QuIET generates models that are human understandable. We describe the method and evaluate it empirically against Support Vector Machines, demonstrating a comparable performance for a known curated dataset and a superior performance for some categories of noisy local businesses data. We also describe an active learning approach that is applicable to QuIET and can boost its performance.
我们提出了一种新的文本二分类算法——QuIET。该方法从一组带注释的文档自动生成一组跨查询,并使用该查询集对未标记的文本进行分类。QuIET生成人类可以理解的模型。我们描述了该方法,并根据支持向量机对其进行了经验评估,展示了对已知策划数据集的可比性能,以及对某些类别的嘈杂本地企业数据的优越性能。我们还描述了一种适用于QuIET的主动学习方法,可以提高其性能。
{"title":"QuIET: A Text Classification Technique Using Automatically Generated Span Queries","authors":"Vassilis Polychronopoulos, N. Pendar, S. Jeffery","doi":"10.1109/ICSC.2014.18","DOIUrl":"https://doi.org/10.1109/ICSC.2014.18","url":null,"abstract":"We propose a novel algorithm, QuIET, for binary classification of texts. The method automatically generates a set of span queries from a set of annotated documents and uses the query set to categorize unlabeled texts. QuIET generates models that are human understandable. We describe the method and evaluate it empirically against Support Vector Machines, demonstrating a comparable performance for a known curated dataset and a superior performance for some categories of noisy local businesses data. We also describe an active learning approach that is applicable to QuIET and can boost its performance.","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125804736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Structure Similarity of Attributed Generalized Trees 带属性广义树的结构相似性
Pub Date : 2014-06-16 DOI: 10.1109/ICSC.2014.33
Mahsa Kiani, V. Bhavsar, H. Boley
Structure-similarity method for attributed generalized trees is proposed. (Meta)data is expressed as a generalized tree, in which inner-vertex labels (as types) and edge labels (as attributes) embody semantic information, while edge weights express assessments regarding the (percentage-)relative importance of the attributes, a kind of pragmatic information added by domain experts. The generalized trees are uniformly represented and interchanged using a weighted extension of Object Oriented RuleML. The recursive similarity algorithm performs a top-down traversal of structures and computes the global similarity of two structures bottom-up considering vertex labels, edge labels, and edge-weight similarities. In order to compare generalized trees having different sizes, the effect of a missing sub-structure on the overall similarity is computed using a simplicity measure. The proposed similarity approach is applied in the retrieval of Electronic Medical Records (EMRs).
提出了带属性广义树的结构相似度方法。(元)数据表示为广义树,其中内顶点标签(作为类型)和边缘标签(作为属性)体现语义信息,而边缘权重表示对属性(百分比-)相对重要性的评估,这是一种由领域专家添加的实用信息。使用面向对象规则ml的加权扩展统一表示和交换广义树。递归相似度算法对结构进行自顶向下的遍历,并考虑顶点标签、边缘标签和边权相似度,自底向上计算两个结构的全局相似度。为了比较具有不同大小的广义树,使用简单度量方法计算缺失子结构对总体相似度的影响。将提出的相似度方法应用于电子病历的检索。
{"title":"Structure Similarity of Attributed Generalized Trees","authors":"Mahsa Kiani, V. Bhavsar, H. Boley","doi":"10.1109/ICSC.2014.33","DOIUrl":"https://doi.org/10.1109/ICSC.2014.33","url":null,"abstract":"Structure-similarity method for attributed generalized trees is proposed. (Meta)data is expressed as a generalized tree, in which inner-vertex labels (as types) and edge labels (as attributes) embody semantic information, while edge weights express assessments regarding the (percentage-)relative importance of the attributes, a kind of pragmatic information added by domain experts. The generalized trees are uniformly represented and interchanged using a weighted extension of Object Oriented RuleML. The recursive similarity algorithm performs a top-down traversal of structures and computes the global similarity of two structures bottom-up considering vertex labels, edge labels, and edge-weight similarities. In order to compare generalized trees having different sizes, the effect of a missing sub-structure on the overall similarity is computed using a simplicity measure. The proposed similarity approach is applied in the retrieval of Electronic Medical Records (EMRs).","PeriodicalId":175352,"journal":{"name":"2014 IEEE International Conference on Semantic Computing","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133531333","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2014 IEEE International Conference on Semantic Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1