Detecting anomalies in the online social network is a significant task as it assists in revealing the useful and interesting information about the user behavior on the network. This paper proposes a rule-based hybrid method using graph theory, Fuzzy clustering and Fuzzy rules for modeling user relationships inherent in online-social-network and for identifying anomalies. Fuzzy C-Means clustering is used to cluster the data and Fuzzy inference engine is used to generate rules based on the cluster behavior. The proposed method is able to achieve improved accuracy for identifying anomalies in comparison to existing methods.
{"title":"A Rule-Based Hybrid Method for Anomaly Detection in Online-Social-Network Graphs","authors":"R. Hassanzadeh, R. Nayak","doi":"10.1109/ICTAI.2013.60","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.60","url":null,"abstract":"Detecting anomalies in the online social network is a significant task as it assists in revealing the useful and interesting information about the user behavior on the network. This paper proposes a rule-based hybrid method using graph theory, Fuzzy clustering and Fuzzy rules for modeling user relationships inherent in online-social-network and for identifying anomalies. Fuzzy C-Means clustering is used to cluster the data and Fuzzy inference engine is used to generate rules based on the cluster behavior. The proposed method is able to achieve improved accuracy for identifying anomalies in comparison to existing methods.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"152 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127379412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Epistemic specification (ES for short) is an extension of answer set programming (ASP for short). The extension is built around the introduction of modalities K and M, and then is capable of representing incomplete information in the presence of multiple belief sets. Although both syntax and semantics of ES are up in the air, the need for this extension has been illustrated with several examples in the literatures. In this paper, we present a new ES version with only modality K and the design of its inference engine ESmodels that aims to be efficient enough to promote the theoretical research and also practical use of ES. We first introduce the syntax and semantics of the new version of ES and show it is succinct but flexible by comparing it with existing ES versions. Then, we focus on the description of the algorithm and optimization approaches of the inference engine. Finally, we conclude with perspectives.
{"title":"ESmodels: An Inference Engine of Epistemic Specifications","authors":"Zhizheng Zhang, Kaikai Zhao, Rongcun Cui","doi":"10.1109/ICTAI.2013.118","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.118","url":null,"abstract":"Epistemic specification (ES for short) is an extension of answer set programming (ASP for short). The extension is built around the introduction of modalities K and M, and then is capable of representing incomplete information in the presence of multiple belief sets. Although both syntax and semantics of ES are up in the air, the need for this extension has been illustrated with several examples in the literatures. In this paper, we present a new ES version with only modality K and the design of its inference engine ESmodels that aims to be efficient enough to promote the theoretical research and also practical use of ES. We first introduce the syntax and semantics of the new version of ES and show it is succinct but flexible by comparing it with existing ES versions. Then, we focus on the description of the algorithm and optimization approaches of the inference engine. Finally, we conclude with perspectives.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127770280","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Several studies in Qualitative Spatial and Temporal Reasoningdiscuss translations of the satisfiability problem on qualitativeconstraint languages into propositional SAT. Most of these encodings focus on compactness, while propagation strength is seldom discussed. In this work, we focus on temporal reasoning with the Point Algebra and Allen's Interval Algebra. We understand all encodings as a combination of propagation andsearch. We first give a systematic analysis of existing propagation approachesfor these constraint languages. They are studied and ordered with respect to their propagation strengthand refutation completeness for classes of input instances. Secondly, we discuss how existing encodings can be derived fromsuch propagation approaches. We conclude our work with an empirical evaluation which shows that theolder ORD-encoding by Nebel and Bürckert performs better than more recently suggested encodings.
{"title":"On the Propagation Strength of SAT Encodings for Qualitative Temporal Reasoning","authors":"Matthias Westphal, J. Hué, S. Wölfl","doi":"10.1109/ICTAI.2013.18","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.18","url":null,"abstract":"Several studies in Qualitative Spatial and Temporal Reasoningdiscuss translations of the satisfiability problem on qualitativeconstraint languages into propositional SAT. Most of these encodings focus on compactness, while propagation strength is seldom discussed. In this work, we focus on temporal reasoning with the Point Algebra and Allen's Interval Algebra. We understand all encodings as a combination of propagation andsearch. We first give a systematic analysis of existing propagation approachesfor these constraint languages. They are studied and ordered with respect to their propagation strengthand refutation completeness for classes of input instances. Secondly, we discuss how existing encodings can be derived fromsuch propagation approaches. We conclude our work with an empirical evaluation which shows that theolder ORD-encoding by Nebel and Bürckert performs better than more recently suggested encodings.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125831724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper presents a simplified approach to performing the Intelligent Water Drops (IWD) process. This approach is designed to be comparatively lightweight while approximating the results of the full IWD process. The Simplified Intelligent Water Drops (SIWD) approach is specifically designed for applications where IWD must be run in a computationally limited environment (such as on a robot, UAV or small spacecraft) or where performance speed must be maximized for time sensitive applications. The SWID approach is described and compared and contracted to the base IWD approach.
{"title":"Characterization of Extended and Simplified Intelligent Water Drop (SIWD) Approaches and Their Comparison to the Intelligent Water Drop (IWD) Approach","authors":"J. Straub, Eunjin Kim","doi":"10.1109/ICTAI.2013.25","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.25","url":null,"abstract":"This paper presents a simplified approach to performing the Intelligent Water Drops (IWD) process. This approach is designed to be comparatively lightweight while approximating the results of the full IWD process. The Simplified Intelligent Water Drops (SIWD) approach is specifically designed for applications where IWD must be run in a computationally limited environment (such as on a robot, UAV or small spacecraft) or where performance speed must be maximized for time sensitive applications. The SWID approach is described and compared and contracted to the base IWD approach.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116798930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sultan Alhusain, S. Coupland, R. John, Maria Kavanagh
Software design patterns describe recurring design problems and provide the essence of best practice solutions. It is useful and important, for various software engineering tasks, to know which design pattern is implemented where in a software design. However, this information is often lost due to poor or absent documentation, and so accurate recognition tools are required. The problem is that design patterns, given their abstract and vague nature, have a level of resistance to be automatically and accurately recognized. Although this vagueness or fuzziness can be captured and modelled by the fuzzy inference system, it has not yet been applied to solve this problem. This paper fills this gap by proposing an approach for design pattern recognition based on Adaptive Neuro Fuzzy Inference System. Our approach consists of two phases: space reduction phase and design pattern recognition phase. Both phases are implemented by ANFIS. We evaluate the approach by an experiment conducted to recognize six design patterns in an open source application. The results show that the approach is viable and promising.
{"title":"Design Pattern Recognition by Using Adaptive Neuro Fuzzy Inference System","authors":"Sultan Alhusain, S. Coupland, R. John, Maria Kavanagh","doi":"10.1109/ICTAI.2013.92","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.92","url":null,"abstract":"Software design patterns describe recurring design problems and provide the essence of best practice solutions. It is useful and important, for various software engineering tasks, to know which design pattern is implemented where in a software design. However, this information is often lost due to poor or absent documentation, and so accurate recognition tools are required. The problem is that design patterns, given their abstract and vague nature, have a level of resistance to be automatically and accurately recognized. Although this vagueness or fuzziness can be captured and modelled by the fuzzy inference system, it has not yet been applied to solve this problem. This paper fills this gap by proposing an approach for design pattern recognition based on Adaptive Neuro Fuzzy Inference System. Our approach consists of two phases: space reduction phase and design pattern recognition phase. Both phases are implemented by ANFIS. We evaluate the approach by an experiment conducted to recognize six design patterns in an open source application. The results show that the approach is viable and promising.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124969912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Javier Barrachina, Piedad Garrido, Manuel Fogué, F. Martinez, Juan-Carlos Cano, C. Calafate, P. Manzoni
A critical issue, especially in urban areas, is the occurrence of traffic accidents, since it could generate traffic jams. Additionally, these traffic jams will negatively affect to the rescue process, increasing the emergency services arrival time, which can determine the difference between life or death for injured people involved in the accident. In this paper, we propose four different approaches addressing the traffic congestion problem, comparing them to obtain the best solution. Using V2I communications, we are able to accurately estimate the traffic density in a certain area, which represents a key parameter to perform efficient traffic redirection, thereby reducing the emergency services arrival time, and avoiding traffic jams when an accident occurs. Specifically, we propose two approaches based on the Dijkstra algorithm, and two approaches based on Evolution Strategies. Results indicate that the Density-Based Evolution Strategy system is the best one among all the proposed solutions, since it offers the lowest emergency services travel times.
{"title":"Using Evolution Strategies to Reduce Emergency Services Arrival Time in Case of Accident","authors":"Javier Barrachina, Piedad Garrido, Manuel Fogué, F. Martinez, Juan-Carlos Cano, C. Calafate, P. Manzoni","doi":"10.1109/ICTAI.2013.127","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.127","url":null,"abstract":"A critical issue, especially in urban areas, is the occurrence of traffic accidents, since it could generate traffic jams. Additionally, these traffic jams will negatively affect to the rescue process, increasing the emergency services arrival time, which can determine the difference between life or death for injured people involved in the accident. In this paper, we propose four different approaches addressing the traffic congestion problem, comparing them to obtain the best solution. Using V2I communications, we are able to accurately estimate the traffic density in a certain area, which represents a key parameter to perform efficient traffic redirection, thereby reducing the emergency services arrival time, and avoiding traffic jams when an accident occurs. Specifically, we propose two approaches based on the Dijkstra algorithm, and two approaches based on Evolution Strategies. Results indicate that the Density-Based Evolution Strategy system is the best one among all the proposed solutions, since it offers the lowest emergency services travel times.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131242418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In a multiplex network, agents are connected by multiple types of links, and the network can be split into more than one network layer which is composed of the same type of links and involved agents. Traditional task allocation methods of multiagent systems only consider the situations of agents themselves, but neglect the effects of network layers in multiplex networks. To solve such a problem, this paper takes network layers into account and presents a novel network layer-oriented task allocation model for multiplex agent networks, with such a model, first the network layers that can satisfy the objectives of task allocation will be allocated, then the final agents will be selected from the allocated network layers. Moreover, this paper deals with the situation of undependable networks, where the resource access of tasks may be undependable, and implements a task allocation based on negotiation reputation. It shows that the network layer-oriented task allocation model leads to an improvement in the success rate and execution time of tasks in multiplex networks when compared to traditional agent-oriented task allocation methods, moreover, such a model has good scalability for the size of tasks and robustness for dynamic undependability.
{"title":"Network Layer-Oriented Task Allocation for Multiagent Systems in Undependable Multiplex Networks","authors":"Yichuan Jiang, Yifeng Zhou, Yunpeng Li","doi":"10.1109/ICTAI.2013.100","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.100","url":null,"abstract":"In a multiplex network, agents are connected by multiple types of links, and the network can be split into more than one network layer which is composed of the same type of links and involved agents. Traditional task allocation methods of multiagent systems only consider the situations of agents themselves, but neglect the effects of network layers in multiplex networks. To solve such a problem, this paper takes network layers into account and presents a novel network layer-oriented task allocation model for multiplex agent networks, with such a model, first the network layers that can satisfy the objectives of task allocation will be allocated, then the final agents will be selected from the allocated network layers. Moreover, this paper deals with the situation of undependable networks, where the resource access of tasks may be undependable, and implements a task allocation based on negotiation reputation. It shows that the network layer-oriented task allocation model leads to an improvement in the success rate and execution time of tasks in multiplex networks when compared to traditional agent-oriented task allocation methods, moreover, such a model has good scalability for the size of tasks and robustness for dynamic undependability.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133765554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Man Zhu, Zhiqiang Gao, Jeff Z. Pan, Yuting Zhao, Ying Xu, Zhibin Quan
Recent years have seen a dramatic growth of semantic web on the data level, but unfortunately not on the schema level, which contains mostly concept hierarchies. The shortage of schemas makes the semantic web data difficult to be used in many semantic web applications, so schemas learning from semantic web data becomes an increasingly pressing issue. In this paper we propose a novel schemas learning approach -BelNet, which combines description logics (DLs) with Bayesian networks. In this way BelNet is capable to understand and capture the semantics of the data on the one hand, and to handle incompleteness during the learning procedure on the other hand. The main contributions of this work are: (i)we introduce the architecture of BelNet, and corresponding lypropose the ontology learning techniques in it, (ii) we compare the experimental results of our approach with the state-of-the-art ontology learning approaches, and provide discussions from different aspects.
{"title":"Ontology Learning from Incomplete Semantic Web Data by BelNet","authors":"Man Zhu, Zhiqiang Gao, Jeff Z. Pan, Yuting Zhao, Ying Xu, Zhibin Quan","doi":"10.1109/ICTAI.2013.117","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.117","url":null,"abstract":"Recent years have seen a dramatic growth of semantic web on the data level, but unfortunately not on the schema level, which contains mostly concept hierarchies. The shortage of schemas makes the semantic web data difficult to be used in many semantic web applications, so schemas learning from semantic web data becomes an increasingly pressing issue. In this paper we propose a novel schemas learning approach -BelNet, which combines description logics (DLs) with Bayesian networks. In this way BelNet is capable to understand and capture the semantics of the data on the one hand, and to handle incompleteness during the learning procedure on the other hand. The main contributions of this work are: (i)we introduce the architecture of BelNet, and corresponding lypropose the ontology learning techniques in it, (ii) we compare the experimental results of our approach with the state-of-the-art ontology learning approaches, and provide discussions from different aspects.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133510477","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
F. Turkmen, S. Foley, B. O’Sullivan, W. M. Fitzgerald, T. Hadzic, Stylianos Basagiannis, M. Boubekeur
Physical access control policies define sets of rulesthat govern people's access to physical resources such asrooms and buildings. While simple decision-precedence can be used to reconcile different rules that result in conflicting access decisions, the presence of rule conflicts and other rule anomalies can make it difficult for a policy-administrator to comprehend and effectively manage complex policies. In this paper we are concerned with discovering conflicts and computing relaxations of access policies in order to eliminate conflicting rule instances. We propose several SAT based encodings in which these rule conflicts and anomalies areexpressed as explanation style problems. Relaxation techniques are in turn used to eliminate these anomalies by recommending what rules have to be revoked or what permissions have to beremoved from which rules. Moreover, we discuss a relaxation strategy that preserves most of the access constraints of theoriginal policy. Finally we provide a preliminary performancestudy of our techniques. Our approach is applicable to access control policies in general.
{"title":"Explanations and Relaxations for Policy Conflicts in Physical Access Control","authors":"F. Turkmen, S. Foley, B. O’Sullivan, W. M. Fitzgerald, T. Hadzic, Stylianos Basagiannis, M. Boubekeur","doi":"10.1109/ICTAI.2013.57","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.57","url":null,"abstract":"Physical access control policies define sets of rulesthat govern people's access to physical resources such asrooms and buildings. While simple decision-precedence can be used to reconcile different rules that result in conflicting access decisions, the presence of rule conflicts and other rule anomalies can make it difficult for a policy-administrator to comprehend and effectively manage complex policies. In this paper we are concerned with discovering conflicts and computing relaxations of access policies in order to eliminate conflicting rule instances. We propose several SAT based encodings in which these rule conflicts and anomalies areexpressed as explanation style problems. Relaxation techniques are in turn used to eliminate these anomalies by recommending what rules have to be revoked or what permissions have to beremoved from which rules. Moreover, we discuss a relaxation strategy that preserves most of the access constraints of theoriginal policy. Finally we provide a preliminary performancestudy of our techniques. Our approach is applicable to access control policies in general.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133726580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Commercial web search engines include a query suggestion module so that given a user's keyword query, alternative suggestions are offered and served as a guide to assist the user in formulating queries which capture his/her intended information need in a quick and simple manner. Majorityof these modules, however, perform an in-depth analysis oflarge query logs and thus (i) their suggestions are mostlybased on queries frequently posted by users and (ii) theirdesign methodologies cannot be applied to make suggestions oncustomized search applications for enterprises for which theirrespective query logs are not large enough or non-existent. To address these design issues, we have developed PQS, aprobabilistic query suggestion module. Unlike its counterparts, PQS is not constrained by the existence of query logs, sinceit solely relies on the availability of user-generated contentfreely accessible online, such as the Wikipedia.org documentcollection, and applies simple, yet effective, probabilistic-andinformation retrieval-based models, i.e., the Multinomial, BigramLanguage, and Vector Space Models, to provide usefuland diverse query suggestions. Empirical studies conductedusing a set of test queries and the feedbacks provided byMechanical Turk appraisers have verified that PQS makesmore useful suggestions than Yahoo! and is almost as goodas Google and Bing based on the relatively small difference inperformance measures achieved by Google and Bing over PQS.
{"title":"A Probabilistic Query Suggestion Approach without Using Query Logs","authors":"M. T. Shaikh, M. S. Pera, Yiu-Kai Ng","doi":"10.1109/ICTAI.2013.99","DOIUrl":"https://doi.org/10.1109/ICTAI.2013.99","url":null,"abstract":"Commercial web search engines include a query suggestion module so that given a user's keyword query, alternative suggestions are offered and served as a guide to assist the user in formulating queries which capture his/her intended information need in a quick and simple manner. Majorityof these modules, however, perform an in-depth analysis oflarge query logs and thus (i) their suggestions are mostlybased on queries frequently posted by users and (ii) theirdesign methodologies cannot be applied to make suggestions oncustomized search applications for enterprises for which theirrespective query logs are not large enough or non-existent. To address these design issues, we have developed PQS, aprobabilistic query suggestion module. Unlike its counterparts, PQS is not constrained by the existence of query logs, sinceit solely relies on the availability of user-generated contentfreely accessible online, such as the Wikipedia.org documentcollection, and applies simple, yet effective, probabilistic-andinformation retrieval-based models, i.e., the Multinomial, BigramLanguage, and Vector Space Models, to provide usefuland diverse query suggestions. Empirical studies conductedusing a set of test queries and the feedbacks provided byMechanical Turk appraisers have verified that PQS makesmore useful suggestions than Yahoo! and is almost as goodas Google and Bing based on the relatively small difference inperformance measures achieved by Google and Bing over PQS.","PeriodicalId":140309,"journal":{"name":"2013 IEEE 25th International Conference on Tools with Artificial Intelligence","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131704651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}