Ting Wang, D. Maynard, Wim Peters, Kalina Bontcheva, H. Cunningham
Creating domain-specific ontologies is one of the main bottlenecks in the development of the semantic Web. Learning an ontology from linguistic resources is helpful to reduce the costs of ontology creation. In this paper, we describe a method to extract the most related concepts from HowNet, a Chinese-English bilingual knowledge dictionary, in order to create a customized ontology for a particular domain. We introduce a new method to measure relatedness (rather than similarity between concepts), which overcomes some of the traditional problems associated with similar concepts being far apart in the hierarchy. Experiments show encouraging results.
{"title":"Extracting a domain ontology from linguistic resource based on relatedness measurements","authors":"Ting Wang, D. Maynard, Wim Peters, Kalina Bontcheva, H. Cunningham","doi":"10.1109/WI.2005.63","DOIUrl":"https://doi.org/10.1109/WI.2005.63","url":null,"abstract":"Creating domain-specific ontologies is one of the main bottlenecks in the development of the semantic Web. Learning an ontology from linguistic resources is helpful to reduce the costs of ontology creation. In this paper, we describe a method to extract the most related concepts from HowNet, a Chinese-English bilingual knowledge dictionary, in order to create a customized ontology for a particular domain. We introduce a new method to measure relatedness (rather than similarity between concepts), which overcomes some of the traditional problems associated with similar concepts being far apart in the hierarchy. Experiments show encouraging results.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116847345","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yuan-chao Li, Xiao-dong Wu, Degan Zhang, Guangping Zeng
In order to realize proactive service, we design and improve relative fuzzy-neural approaches. Generally, the network can be classified into two. One is that fuzzy logic reasoning is completed by fuzzy weight in neural system. The other is that the input data must be fuzzified in the first or second level, but not weight. We discuss and study the second sort fuzzification in this paper. For proactive decision, fusion method based on fuzzy-neural can make Web-based intelligent system keep advantage of fuzzy logic system and remain adaptive optimum in proactive/attentive service. The correctness and validity of our new approach have been tested.
{"title":"Fuzzy-neural theory applied to Web-based proactive service","authors":"Yuan-chao Li, Xiao-dong Wu, Degan Zhang, Guangping Zeng","doi":"10.1109/WI.2005.65","DOIUrl":"https://doi.org/10.1109/WI.2005.65","url":null,"abstract":"In order to realize proactive service, we design and improve relative fuzzy-neural approaches. Generally, the network can be classified into two. One is that fuzzy logic reasoning is completed by fuzzy weight in neural system. The other is that the input data must be fuzzified in the first or second level, but not weight. We discuss and study the second sort fuzzification in this paper. For proactive decision, fusion method based on fuzzy-neural can make Web-based intelligent system keep advantage of fuzzy logic system and remain adaptive optimum in proactive/attentive service. The correctness and validity of our new approach have been tested.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123202131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Generally, OLAP analysis are based on both the observed data and a set of OLAP operators for restructuration and granularity modification. The goal is to discover patterns hidden into data. Unfortunately, this approach is also based on the analyst background. This latter assumes hypothesis according to his background and analyses data consequently: "hypothesis driven analysis". The integration of knowledge into data warehouse conduce to enriched analysis context where objects and their relations are explicitly represented, handled and visualized. We investigate a deep integration where the basic datawarehouse's operators consider both data and knowledge. This paper applies knowledge datawarehouse concept to Web usage analysis.
{"title":"Knowledge datawarehouse: Web usage OLAP application","authors":"M. Quafafou, S. Naouali, G. Nachouki","doi":"10.1109/WI.2005.88","DOIUrl":"https://doi.org/10.1109/WI.2005.88","url":null,"abstract":"Generally, OLAP analysis are based on both the observed data and a set of OLAP operators for restructuration and granularity modification. The goal is to discover patterns hidden into data. Unfortunately, this approach is also based on the analyst background. This latter assumes hypothesis according to his background and analyses data consequently: \"hypothesis driven analysis\". The integration of knowledge into data warehouse conduce to enriched analysis context where objects and their relations are explicitly represented, handled and visualized. We investigate a deep integration where the basic datawarehouse's operators consider both data and knowledge. This paper applies knowledge datawarehouse concept to Web usage analysis.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128942293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Navigation and interaction patterns of Web users can be relatively complex, especially for sites with interactive applications that support user sessions and profiles. We describe such a case for an interactive virtual garment dressing room. The application is distributed over many Web sites, supports personalization and user profiles, and the notion of a multi-site user session. It has its own data logging system that generates approximately 5GB of complex data per month. The analysis of those logs requires more sophisticated processing than is typically done using a relational language. Even the use of procedural languages and DBMS can prove tedious and inefficient. We show an approach to the analysis of complex log data based on a parallel stream processing architecture and the use of specialized languages, namely a grammatical parser and a logic programming module that offers an efficient, flexible, and powerful solution.
{"title":"Web log session analyzer: integrating parsing and logic programming into a data mart architecture","authors":"M. Desmarais","doi":"10.1109/WI.2005.159","DOIUrl":"https://doi.org/10.1109/WI.2005.159","url":null,"abstract":"Navigation and interaction patterns of Web users can be relatively complex, especially for sites with interactive applications that support user sessions and profiles. We describe such a case for an interactive virtual garment dressing room. The application is distributed over many Web sites, supports personalization and user profiles, and the notion of a multi-site user session. It has its own data logging system that generates approximately 5GB of complex data per month. The analysis of those logs requires more sophisticated processing than is typically done using a relational language. Even the use of procedural languages and DBMS can prove tedious and inefficient. We show an approach to the analysis of complex log data based on a parallel stream processing architecture and the use of specialized languages, namely a grammatical parser and a logic programming module that offers an efficient, flexible, and powerful solution.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130083543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A-globe (Speretta and Gauch, 2005) is a simulation oriented multi-agent platform featuring agent migration, communication inaccessibility simulation and high scalability with moderate hardware requirements. Using dedicated simulation messaging together with 2D and 3D visualization support, large agent systems can be engineered, tested and visualized on a single machine. A-globe agents are fully fledged JAVA agents, each with its own independent thread that can autonomously migrate between platforms running on different hosts. Thanks to the separation of simulation and agent code, deployment of agents to embedded devices is straightforward. Platform is not natively FIPA-compliant, as the interoperability was sacrificed to support the scalability and efficiency.
a -globe (Speretta and Gauch, 2005)是一个面向仿真的多智能体平台,具有智能体迁移、通信不可达性仿真和高可扩展性,硬件要求适中。使用专用的仿真消息传递以及2D和3D可视化支持,可以在一台机器上设计、测试和可视化大型代理系统。全局代理是完全成熟的JAVA代理,每个代理都有自己独立的线程,可以在不同主机上运行的平台之间自主迁移。由于仿真和代理代码的分离,将代理部署到嵌入式设备非常简单。平台本身不符合fipa,因为为了支持可伸缩性和效率而牺牲了互操作性。
{"title":"A-globe: multi-agent platform with advanced simulation and visualization support","authors":"D. Sislák, M. Rehák, M. Pechoucek","doi":"10.1109/WI.2005.18","DOIUrl":"https://doi.org/10.1109/WI.2005.18","url":null,"abstract":"A-globe (Speretta and Gauch, 2005) is a simulation oriented multi-agent platform featuring agent migration, communication inaccessibility simulation and high scalability with moderate hardware requirements. Using dedicated simulation messaging together with 2D and 3D visualization support, large agent systems can be engineered, tested and visualized on a single machine. A-globe agents are fully fledged JAVA agents, each with its own independent thread that can autonomously migrate between platforms running on different hosts. Thanks to the separation of simulation and agent code, deployment of agents to embedded devices is straightforward. Platform is not natively FIPA-compliant, as the interoperability was sacrificed to support the scalability and efficiency.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"146 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127513359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Labský, V. Svátek, Ondrej Sváb-Zamazal, P. Praks, M. Krátký, V. Snás̃el
We describe an application of information extraction from company Web sites focusing on product offers. A statistical approach to text analysis is used in conjunction with different ways of image classification. Ontological knowledge is used to group the extracted items into structured objects. The results are stored in an RDF repository and made available for structured search.
{"title":"Information extraction from HTML product catalogues: from source code and images to RDF","authors":"M. Labský, V. Svátek, Ondrej Sváb-Zamazal, P. Praks, M. Krátký, V. Snás̃el","doi":"10.1109/WI.2005.78","DOIUrl":"https://doi.org/10.1109/WI.2005.78","url":null,"abstract":"We describe an application of information extraction from company Web sites focusing on product offers. A statistical approach to text analysis is used in conjunction with different ways of image classification. Ontological knowledge is used to group the extracted items into structured objects. The results are stored in an RDF repository and made available for structured search.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129967369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
General tensors can represent colour images more naturally than conventional features; however, the general tensors' stability properties are not reported and remain to be a key problem. In this paper, we use the tensor minimax probability (TMPM) to prove that the tensor representation is stable. The proof is based on the random subspace method through a large number of experiments.
{"title":"Stable third-order tensor representation for colour image classification","authors":"D. Tao, S. Maybank, Weiming Hu, Xuelong Li","doi":"10.1109/WI.2005.136","DOIUrl":"https://doi.org/10.1109/WI.2005.136","url":null,"abstract":"General tensors can represent colour images more naturally than conventional features; however, the general tensors' stability properties are not reported and remain to be a key problem. In this paper, we use the tensor minimax probability (TMPM) to prove that the tensor representation is stable. The proof is based on the random subspace method through a large number of experiments.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120940508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Zongxia Du, J. Huai, Yunhao Liu, Chunming Hu, Lei Lei
Inter-organizational business processes usually require complex and time-consuming interactions between partners than simple interactions supported by WSDL. Automated reconciliation is essential to enable dynamic inter-organizational business collaboration. To the best of our knowledge, however, there is not a practical automated reconciliation algorithm available. In this paper, we propose a practical automated reconciliation algorithm, called IPR (interaction process reconciliation) based on Petri net, which is able to effectively facilitate dynamic interactions among trading partners in a peer-to-peer fashion. We implement a prototype IPR server in our lab, and evaluate our design by comprehensive experiments. Results show that IPR significantly outperforms existing approaches in terms of matching success rate, response time, and matching efficiency.
{"title":"IPR: automated interaction process reconciliation","authors":"Zongxia Du, J. Huai, Yunhao Liu, Chunming Hu, Lei Lei","doi":"10.1109/WI.2005.84","DOIUrl":"https://doi.org/10.1109/WI.2005.84","url":null,"abstract":"Inter-organizational business processes usually require complex and time-consuming interactions between partners than simple interactions supported by WSDL. Automated reconciliation is essential to enable dynamic inter-organizational business collaboration. To the best of our knowledge, however, there is not a practical automated reconciliation algorithm available. In this paper, we propose a practical automated reconciliation algorithm, called IPR (interaction process reconciliation) based on Petri net, which is able to effectively facilitate dynamic interactions among trading partners in a peer-to-peer fashion. We implement a prototype IPR server in our lab, and evaluate our design by comprehensive experiments. Results show that IPR significantly outperforms existing approaches in terms of matching success rate, response time, and matching efficiency.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126642677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mani Abrol, Bhavin Doshi, Jim Kanihan, Amit Kumar, Jinhui Liu, Jianchang Mao
The taxonomy is the most popular way for organizing a large volume of content. A taxonomy, which is typically a hierarchical representation of categories, provides a navigation structure for exploring and understanding the underlying corpus without sifting through a huge volume of documents. Creating and maintaining taxonomies with a large volume of documents remains a daunting task facing many enterprises. Content organization process typically involves four basic steps: (i) creating taxonomies; (ii) building classification models; (iii) populating taxonomies with documents; and (iv) deploying populated taxonomies in enterprise portals. Each step in the process may have unique requirements that determine what techniques and tools are suitable for the tasks. In this paper, we present a comprehensive suite of tools developed by Verity Inc. for accurately, collaboratively, and efficiently organizing enterprise content.
{"title":"Intelligent taxonomy management tools for enterprise content","authors":"Mani Abrol, Bhavin Doshi, Jim Kanihan, Amit Kumar, Jinhui Liu, Jianchang Mao","doi":"10.1109/WI.2005.82","DOIUrl":"https://doi.org/10.1109/WI.2005.82","url":null,"abstract":"The taxonomy is the most popular way for organizing a large volume of content. A taxonomy, which is typically a hierarchical representation of categories, provides a navigation structure for exploring and understanding the underlying corpus without sifting through a huge volume of documents. Creating and maintaining taxonomies with a large volume of documents remains a daunting task facing many enterprises. Content organization process typically involves four basic steps: (i) creating taxonomies; (ii) building classification models; (iii) populating taxonomies with documents; and (iv) deploying populated taxonomies in enterprise portals. Each step in the process may have unique requirements that determine what techniques and tools are suitable for the tasks. In this paper, we present a comprehensive suite of tools developed by Verity Inc. for accurately, collaboratively, and efficiently organizing enterprise content.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"168 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126744257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hongbo Zou, Hai Jin, Zongfen Han, J. Tie, Xuanhua Shi
Resource discovery is a basic service in grid computing: gives a description of resources desired and finds the available one to match the description. In computational grid, how to discover resources efficiently has become a crucial factor to evaluate the performance in the whole system. In this paper, we present a bid-based resource discovery algorithm, which converts a resource request into a bidding letter and sends it to a group of physical services owned by the same virtual service to call for bidding. All resources receiving bidding letter make offers to bid according to our algorithm. Job manager selects the best one to response client request. To evaluate the performance of our method, we compare our system with the centralized and peer-to-peer resource discovery approaches. The analysis results show that our system reduces average response time of jobs, leverages the cost of the resource discovery, and improves the system scalability.
{"title":"A virtual-service-domain based bidding algorithm for resource discovery in computational grid","authors":"Hongbo Zou, Hai Jin, Zongfen Han, J. Tie, Xuanhua Shi","doi":"10.1109/WI.2005.17","DOIUrl":"https://doi.org/10.1109/WI.2005.17","url":null,"abstract":"Resource discovery is a basic service in grid computing: gives a description of resources desired and finds the available one to match the description. In computational grid, how to discover resources efficiently has become a crucial factor to evaluate the performance in the whole system. In this paper, we present a bid-based resource discovery algorithm, which converts a resource request into a bidding letter and sends it to a group of physical services owned by the same virtual service to call for bidding. All resources receiving bidding letter make offers to bid according to our algorithm. Job manager selects the best one to response client request. To evaluate the performance of our method, we compare our system with the centralized and peer-to-peer resource discovery approaches. The analysis results show that our system reduces average response time of jobs, leverages the cost of the resource discovery, and improves the system scalability.","PeriodicalId":213856,"journal":{"name":"The 2005 IEEE/WIC/ACM International Conference on Web Intelligence (WI'05)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125856247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}