首页 > 最新文献

Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics最新文献

英文 中文
Addressing the cold user problem for model-based recommender systems 解决基于模型的推荐系统的冷用户问题
Tomas Geurts, F. Frasincar
Customers of a webshop are often presented large assortments, which can lead to customers struggling finding their desired product(s), an issue known as choice overload. In order to overcome this issue, recommender systems are used in webshops to provide personalized product recommendations to customers. Though, recommender systems using matrix factorization are not able to provide recommendations to new customers (i.e., cold users). To facilitate recommendations to cold users we investigate multiple active learning strategies, and subsequently evaluate which active learning strategy is able to optimally elicit the preferences from the cold users. Our model is empirically validated using a dataset from the webshop of de Bijenkorf, a Dutch department store. We find that the overall best-performing active learning strategy is PopGini, an active learning strategy which combines the popularity of an item with its Gini impurity score.
网上商店的顾客经常会看到大量的分类,这可能会导致顾客很难找到他们想要的产品,这个问题被称为选择过载。为了克服这个问题,在网上商店中使用推荐系统向客户提供个性化的产品推荐。然而,使用矩阵分解的推荐系统无法向新客户(即冷用户)提供推荐。为了便于向冷用户推荐,我们研究了多种主动学习策略,并随后评估了哪种主动学习策略能够最优地引起冷用户的偏好。我们的模型使用来自荷兰百货公司de Bijenkorf网店的数据集进行了实证验证。我们发现,总体上表现最好的主动学习策略是PopGini,这是一种将项目的受欢迎程度与其基尼杂质分数相结合的主动学习策略。
{"title":"Addressing the cold user problem for model-based recommender systems","authors":"Tomas Geurts, F. Frasincar","doi":"10.1145/3106426.3106431","DOIUrl":"https://doi.org/10.1145/3106426.3106431","url":null,"abstract":"Customers of a webshop are often presented large assortments, which can lead to customers struggling finding their desired product(s), an issue known as choice overload. In order to overcome this issue, recommender systems are used in webshops to provide personalized product recommendations to customers. Though, recommender systems using matrix factorization are not able to provide recommendations to new customers (i.e., cold users). To facilitate recommendations to cold users we investigate multiple active learning strategies, and subsequently evaluate which active learning strategy is able to optimally elicit the preferences from the cold users. Our model is empirically validated using a dataset from the webshop of de Bijenkorf, a Dutch department store. We find that the overall best-performing active learning strategy is PopGini, an active learning strategy which combines the popularity of an item with its Gini impurity score.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77545667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Ontology of human relation extraction based on dependency syntax rules 基于依赖语法规则的人际关系抽取本体
Long He, Likun Qiu
This paper proposed a novel scheme for extracting character relation from unstructured text based on dependency grammar rules. First of all, we took the Three Kingdoms characters as our research object, then selected articles containing target relationships and thus constructed a corpus consisting of 1000 sentences. Secondly, We analyzed the corpus and developed a set of dependent grammar rules for relation extraction. Finally, we proposed a system, which makes it possible for computers to automatically extract and identify character relationships.
提出了一种基于依赖语法规则的非结构化文本字符关系提取方法。首先,我们以三国文字为研究对象,选取含有目标关系的文章,构建了一个1000句的语料库。其次,我们对语料库进行了分析,并开发了一套用于关系抽取的依赖语法规则。最后,我们提出了一个系统,使计算机能够自动提取和识别字符关系。
{"title":"Ontology of human relation extraction based on dependency syntax rules","authors":"Long He, Likun Qiu","doi":"10.1145/3106426.3109050","DOIUrl":"https://doi.org/10.1145/3106426.3109050","url":null,"abstract":"This paper proposed a novel scheme for extracting character relation from unstructured text based on dependency grammar rules. First of all, we took the Three Kingdoms characters as our research object, then selected articles containing target relationships and thus constructed a corpus consisting of 1000 sentences. Secondly, We analyzed the corpus and developed a set of dependent grammar rules for relation extraction. Finally, we proposed a system, which makes it possible for computers to automatically extract and identify character relationships.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76215119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Augmented SVM with ordinal partitioning for text classification 基于有序划分的增强支持向量机文本分类
Yong Shi, Peijia Li, Lingfeng Niu
Ordinal regression has received increasing interest in the past years. It aims to classify patterns by an ordinal scale. With the the explosive growth of data, the method of SVM with ordinal partitioning called SVMOP highlights its advantages due to its convenience of dealing with large scale data. However, the method of SVMOP for ordinal regression has not been exploited much. As we know, the costs should be different when dealing with mislabeled samples and how to use them plays a dominant role in model building. However, L2-loss which could enlarge the cost sensitivity has not been applied into SVM ordinal partition yet. In this paper, we propose the method of SVMOP with L2-loss for ordinal regression. Numerical results show that our approach outperforms the method of SVMOP with L1-loss and other ordianl regression models.
序数回归在过去几年中受到越来越多的关注。它的目的是按顺序对模式进行分类。随着数据的爆炸式增长,被称为SVMOP的有序划分支持向量机方法因其处理大规模数据的便捷性而凸显出其优势。然而,用于有序回归的SVMOP方法并没有得到太多的应用。正如我们所知,处理错标样本的成本应该是不同的,如何使用它们在模型构建中起着主导作用。而L2-loss会增大代价敏感性,目前还没有应用到支持向量机的有序划分中。本文提出了带L2-loss的SVMOP有序回归方法。数值结果表明,该方法优于具有L1-loss的SVMOP方法和其他正常回归模型。
{"title":"Augmented SVM with ordinal partitioning for text classification","authors":"Yong Shi, Peijia Li, Lingfeng Niu","doi":"10.1145/3106426.3109428","DOIUrl":"https://doi.org/10.1145/3106426.3109428","url":null,"abstract":"Ordinal regression has received increasing interest in the past years. It aims to classify patterns by an ordinal scale. With the the explosive growth of data, the method of SVM with ordinal partitioning called SVMOP highlights its advantages due to its convenience of dealing with large scale data. However, the method of SVMOP for ordinal regression has not been exploited much. As we know, the costs should be different when dealing with mislabeled samples and how to use them plays a dominant role in model building. However, L2-loss which could enlarge the cost sensitivity has not been applied into SVM ordinal partition yet. In this paper, we propose the method of SVMOP with L2-loss for ordinal regression. Numerical results show that our approach outperforms the method of SVMOP with L1-loss and other ordianl regression models.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74502104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Constructing and visualizing topic forests for text streams 为文本流构建和可视化主题森林
Takayasu Fushimi, T. Satoh
A great deal of such texts as news and blog articles, web pages, and scientific literature are posted on the web as time goes by, and are generally called time-series documents or text streams. For each document, some strongly or weakly relevant texts exist. Although such relevance is represented as citations among scientific literatures, trackback among blog articles, hyperlinks among Wikipedia articles or web pages and so on, the relevance among news articles is not always clearly specified. One easy way to build a similarity network is by calculating the similarity among news articles and making links among similar articles; however, adding information about the posted times of articles to a similarity network is difficult. To overcome this problem, we propose a framework that consists of two parts: 1) tree structures called Topic Forests and 2) their visualization. Topic Forests are constructed by semantically and temporally linking cohesive texts while preserving their posted order. We provide effective access for users to text streams by embedding Topic Forests over the polar coordinates with a technique called Polar Coordinate Embedding. From experimental evaluations using the actual text streams of news articles, we confirm that Topic Forests semantically and temporally maintain cohesiveness, and Polar Coordinate Embedding achieves effective accessibility.
大量的文本,如新闻和博客文章、网页和科学文献,随着时间的推移被发布在网络上,通常被称为时间序列文档或文本流。对于每个文档,存在一些强相关或弱相关的文本。虽然这种相关性表现为科学文献之间的引用、博客文章之间的追溯、维基百科文章或网页之间的超链接等,但新闻文章之间的相关性并不总是明确规定的。建立相似网络的一个简单方法是计算新闻文章之间的相似度,并在相似的文章之间建立链接;然而,在相似网络中添加关于文章发布时间的信息是困难的。为了克服这个问题,我们提出了一个由两部分组成的框架:1)称为主题森林的树结构和2)它们的可视化。主题森林是通过语义上和时间上连接内聚文本,同时保持其发布顺序来构建的。我们使用一种称为极坐标嵌入的技术在极坐标上嵌入主题森林,为用户提供对文本流的有效访问。通过对实际新闻文本流的实验评估,我们证实了主题森林在语义和时间上保持了内聚性,极坐标嵌入实现了有效的可达性。
{"title":"Constructing and visualizing topic forests for text streams","authors":"Takayasu Fushimi, T. Satoh","doi":"10.1145/3106426.3106455","DOIUrl":"https://doi.org/10.1145/3106426.3106455","url":null,"abstract":"A great deal of such texts as news and blog articles, web pages, and scientific literature are posted on the web as time goes by, and are generally called time-series documents or text streams. For each document, some strongly or weakly relevant texts exist. Although such relevance is represented as citations among scientific literatures, trackback among blog articles, hyperlinks among Wikipedia articles or web pages and so on, the relevance among news articles is not always clearly specified. One easy way to build a similarity network is by calculating the similarity among news articles and making links among similar articles; however, adding information about the posted times of articles to a similarity network is difficult. To overcome this problem, we propose a framework that consists of two parts: 1) tree structures called Topic Forests and 2) their visualization. Topic Forests are constructed by semantically and temporally linking cohesive texts while preserving their posted order. We provide effective access for users to text streams by embedding Topic Forests over the polar coordinates with a technique called Polar Coordinate Embedding. From experimental evaluations using the actual text streams of news articles, we confirm that Topic Forests semantically and temporally maintain cohesiveness, and Polar Coordinate Embedding achieves effective accessibility.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84991192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automated classification of EEG signals for predicting students' cognitive state during learning 脑电信号自动分类预测学生学习过程中的认知状态
Xi Liu, P. Tan, Lei Liu, S. Simske
For distance learning applications, inferring the cognitive states of students, particularly, their concentration and comprehension levels during instruction, is important to assess their learning efficacy. In this paper, we investigated the feasibility of using EEG recordings generated from an off-the-shelf, wearable device to automatically classify the cognitive states of students as they were asked to perform a series of reading and question answering tasks. We showed that the EEG data can effectively predict whether a student is attentive or distracted as well as the student's reading speed, which is an important measure of reading fluency. However, the EEG signals alone are insufficient to predict how well the students can correctly answer questions related to the reading materials as there were other confounding factors, such as the students' background knowledge, that must be taken into consideration. We also showed that the accuracy in predicting the different cognitive states depends on the choice of classifier used (global, local, or multi-task learning). For example, the concentration level of a student can be accurately predicted using a local model whereas a global model that incorporates side information about the student's background knowledge is more effective at predicting whether the student will correctly answer questions about the materials they read.
在远程学习应用中,推断学生的认知状态,特别是他们在教学中的注意力和理解水平,对于评估他们的学习效果是重要的。在本文中,我们研究了使用现成的可穿戴设备生成的脑电图记录来自动分类学生在执行一系列阅读和问答任务时的认知状态的可行性。我们发现脑电图数据可以有效地预测学生的注意力是否集中,以及学生的阅读速度,这是阅读流畅性的重要衡量标准。然而,仅凭脑电图信号不足以预测学生正确回答阅读材料相关问题的能力,因为还需要考虑学生的背景知识等其他干扰因素。我们还表明,预测不同认知状态的准确性取决于所使用分类器的选择(全局、局部或多任务学习)。例如,使用局部模型可以准确地预测学生的集中程度,而使用包含学生背景知识的侧面信息的全局模型则可以更有效地预测学生是否会正确回答有关他们所阅读材料的问题。
{"title":"Automated classification of EEG signals for predicting students' cognitive state during learning","authors":"Xi Liu, P. Tan, Lei Liu, S. Simske","doi":"10.1145/3106426.3106453","DOIUrl":"https://doi.org/10.1145/3106426.3106453","url":null,"abstract":"For distance learning applications, inferring the cognitive states of students, particularly, their concentration and comprehension levels during instruction, is important to assess their learning efficacy. In this paper, we investigated the feasibility of using EEG recordings generated from an off-the-shelf, wearable device to automatically classify the cognitive states of students as they were asked to perform a series of reading and question answering tasks. We showed that the EEG data can effectively predict whether a student is attentive or distracted as well as the student's reading speed, which is an important measure of reading fluency. However, the EEG signals alone are insufficient to predict how well the students can correctly answer questions related to the reading materials as there were other confounding factors, such as the students' background knowledge, that must be taken into consideration. We also showed that the accuracy in predicting the different cognitive states depends on the choice of classifier used (global, local, or multi-task learning). For example, the concentration level of a student can be accurately predicted using a local model whereas a global model that incorporates side information about the student's background knowledge is more effective at predicting whether the student will correctly answer questions about the materials they read.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84865331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
CrimeProfiler: crime information extraction and visualization from news media CrimeProfiler:从新闻媒体中提取和可视化犯罪信息
Tirthankar Dasgupta, Abir Naskar, Rupsa Saha, Lipika Dey
News articles from different sources regularly report crime incidents that contain details of crime, information about accused entities, details of the investigation process and finally details of judgement. In this paper, we have proposed natural language processing techniques for extraction and curation of crime-related information from digitally published News articles. We have leveraged computational linguistics based methods to analyse crime related News documents to extract different crime related entities and events. This includes name of the criminal, name of the victim, nature of crime, geographic location, date and time, and action taken against the criminal. We have also proposed a semi-supervised learning technique to learn different categories of crime events from the News documents. This helps in continuous evolution of the crime dictionaries. Thus the proposed methods are not restricted to detecting known crimes only but contribute actively towards maintaining an updated crime dictionary. We have done experiments with a collection of 3000 crime-reporting News articles. The end-product of our experiments is a crime-register that contains details of crime committed across geographies and time. This register can be further utilized for analytical and reporting purposes.
来自不同来源的新闻文章定期报道犯罪事件,其中包括犯罪细节、被指控实体的信息、调查过程的细节以及最后的判决细节。在本文中,我们提出了自然语言处理技术,用于从数字发布的新闻文章中提取和管理与犯罪相关的信息。我们利用基于计算语言学的方法来分析犯罪相关的新闻文档,以提取不同的犯罪相关实体和事件。这包括罪犯的姓名、受害者的姓名、犯罪性质、地理位置、日期和时间以及对罪犯采取的行动。我们还提出了一种半监督学习技术,从新闻文档中学习不同类别的犯罪事件。这有助于犯罪词典的不断发展。因此,所提出的方法不仅限于侦查已知的犯罪,而且积极地有助于维护更新的犯罪词典。我们用3000篇犯罪报道的新闻文章做了实验。我们实验的最终成果是一个犯罪登记簿,其中包含了跨越地域和时间的犯罪细节。该登记册可进一步用于分析和报告目的。
{"title":"CrimeProfiler: crime information extraction and visualization from news media","authors":"Tirthankar Dasgupta, Abir Naskar, Rupsa Saha, Lipika Dey","doi":"10.1145/3106426.3106476","DOIUrl":"https://doi.org/10.1145/3106426.3106476","url":null,"abstract":"News articles from different sources regularly report crime incidents that contain details of crime, information about accused entities, details of the investigation process and finally details of judgement. In this paper, we have proposed natural language processing techniques for extraction and curation of crime-related information from digitally published News articles. We have leveraged computational linguistics based methods to analyse crime related News documents to extract different crime related entities and events. This includes name of the criminal, name of the victim, nature of crime, geographic location, date and time, and action taken against the criminal. We have also proposed a semi-supervised learning technique to learn different categories of crime events from the News documents. This helps in continuous evolution of the crime dictionaries. Thus the proposed methods are not restricted to detecting known crimes only but contribute actively towards maintaining an updated crime dictionary. We have done experiments with a collection of 3000 crime-reporting News articles. The end-product of our experiments is a crime-register that contains details of crime committed across geographies and time. This register can be further utilized for analytical and reporting purposes.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73510322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Identifying active, reactive, and inactive targets of socialbots in Twitter 识别活跃的,被动的和不活跃的目标在Twitter上的社交机器人
Mohd Fazil, M. Abulaish
Online social networks are facing serious threats due to presence of human-behaviour imitating malicious bots (aka socialbots) that are successful mainly due to existence of their duped followers. In this paper, we propose an approach to categorize Twitter users into three groups - active, reactive, and inactive targets, based on their interaction behaviour with socialbots. Active users are those who themselves follow socialbots without being followed by them, reactive users respond to the following socialbots by following them back, whereas inactive users do not show any interest against the following requests from anonymous socialbots. The proposed approach is modelled as both binary and ternary classification problem, wherein users' profile is generated using static and dynamic components representing their identical and behavioural aspects. Three different classification techniques viz Naive Bayes, Reduced Error Pruned Decision Tree, and Random Forest are used over a dataset of 749 users collected through live experiment, and a thorough analyses of the identified users categories is presented, wherein it is found that active and reactive users keep on frequently updating their tweets containing advertising related contents. Finally, feature ranking algorithms are used to rank identified features to analyse their discriminative power, and it is found that following rate and follower rate are the most dominating features.
由于模仿人类行为的恶意机器人(又名社交机器人)的存在,在线社交网络正面临着严重的威胁,这些恶意机器人的成功主要是因为它们被欺骗的追随者的存在。在本文中,我们提出了一种方法,根据Twitter用户与社交机器人的互动行为,将Twitter用户分为三组——活跃的、被动的和不活跃的目标。活跃用户是那些自己关注社交机器人而不被他们关注的用户,被动用户通过关注他们来回应以下社交机器人,而不活跃用户对匿名社交机器人的以下请求没有任何兴趣。所提出的方法建模为二元和三元分类问题,其中使用表示其相同和行为方面的静态和动态组件生成用户配置文件。对现场实验收集的749个用户数据集使用了朴素贝叶斯、减少错误修剪决策树和随机森林三种不同的分类技术,并对识别出的用户类别进行了深入分析,发现活跃用户和被动用户都在频繁更新包含广告相关内容的推文。最后,利用特征排序算法对识别出的特征进行排序,分析其判别能力,发现跟随率和跟随率是最主要的特征。
{"title":"Identifying active, reactive, and inactive targets of socialbots in Twitter","authors":"Mohd Fazil, M. Abulaish","doi":"10.1145/3106426.3106483","DOIUrl":"https://doi.org/10.1145/3106426.3106483","url":null,"abstract":"Online social networks are facing serious threats due to presence of human-behaviour imitating malicious bots (aka socialbots) that are successful mainly due to existence of their duped followers. In this paper, we propose an approach to categorize Twitter users into three groups - active, reactive, and inactive targets, based on their interaction behaviour with socialbots. Active users are those who themselves follow socialbots without being followed by them, reactive users respond to the following socialbots by following them back, whereas inactive users do not show any interest against the following requests from anonymous socialbots. The proposed approach is modelled as both binary and ternary classification problem, wherein users' profile is generated using static and dynamic components representing their identical and behavioural aspects. Three different classification techniques viz Naive Bayes, Reduced Error Pruned Decision Tree, and Random Forest are used over a dataset of 749 users collected through live experiment, and a thorough analyses of the identified users categories is presented, wherein it is found that active and reactive users keep on frequently updating their tweets containing advertising related contents. Finally, feature ranking algorithms are used to rank identified features to analyse their discriminative power, and it is found that following rate and follower rate are the most dominating features.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74951068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Current location-based next POI recommendation 基于当前位置的下一个POI推荐
Shokirkhon Oppokhonov, Seyoung Park, Isaac. K. E. Ampomah
Availability of large volume of community contributed location data enables a lot of location providing services and these services have attracted many industries and academic researchers by its importance. In this paper we propose the new recommender system that recommends the new POI for next hours. First we find the users with similar check-in sequences and depict their check-in sequences as a directed graph, then find the users current location. To recommend the new POI recommendation for next hour we refer to the directed graph we have created. Our algorithm considers both the temporal factor i.e., recommendation time, and the spatial(distance) at the same time. We conduct an experiment on random data collected from Foursquare and Gowalla. Experiment results show that our proposed model outperforms the collaborative-filtering based state-of-the-art recommender techniques.
大量社区贡献的位置数据的可用性使得大量的位置提供服务成为可能,这些服务的重要性吸引了许多行业和学术界的研究人员。在本文中,我们提出了一个新的推荐系统,可以为接下来的几个小时推荐新的POI。首先,我们找到具有相似签入序列的用户,并将其签入序列描述为有向图,然后找到用户的当前位置。为了推荐下一个小时的新POI建议,我们参考我们创建的有向图。我们的算法同时考虑了时间因素(即推荐时间)和空间因素(即距离)。我们对从Foursquare和Gowalla收集的随机数据进行了实验。实验结果表明,我们提出的模型优于基于协同过滤的最先进推荐技术。
{"title":"Current location-based next POI recommendation","authors":"Shokirkhon Oppokhonov, Seyoung Park, Isaac. K. E. Ampomah","doi":"10.1145/3106426.3106528","DOIUrl":"https://doi.org/10.1145/3106426.3106528","url":null,"abstract":"Availability of large volume of community contributed location data enables a lot of location providing services and these services have attracted many industries and academic researchers by its importance. In this paper we propose the new recommender system that recommends the new POI for next hours. First we find the users with similar check-in sequences and depict their check-in sequences as a directed graph, then find the users current location. To recommend the new POI recommendation for next hour we refer to the directed graph we have created. Our algorithm considers both the temporal factor i.e., recommendation time, and the spatial(distance) at the same time. We conduct an experiment on random data collected from Foursquare and Gowalla. Experiment results show that our proposed model outperforms the collaborative-filtering based state-of-the-art recommender techniques.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77535947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Sentiment diversification for short review summarization 情绪多元化短评总结
Mohammed Al-Dhelaan, A. Al-Suhaim
With the abundance of reviews published on the Web about a given product, consumers are looking for ways to view major opinions that can be presented in a quick and succinct way. Reviews contain many different opinions, making the ability to show a diversified review summary that focus on coverage and diversity a major goal. Most review summarization work focuses on showing salient reviews as a summary which might ignore diversity in summaries. In this paper, we present a graph-based algorithm that is capable of producing extractive summaries that are both diversified from a sentiment point of view and topically well-covered. First, we use statistical measures to find topical words. Then we split the dataset based on the sentiment class of the reviews and perform the ranking on each sentiment graph. When compared with different baselines, our approach scores best in most ROUGE metrics. Specifically, our approach shows improvements of 3.9% in ROUGE-1 and 1.8% in ROUGE-L in comparison with the best competing baseline.
随着网络上发布的关于某一产品的大量评论,消费者正在寻找能够以快速、简洁的方式显示主要意见的方法。评审包含了许多不同的意见,这使得能够显示一个专注于覆盖范围和多样性的多样化评审总结成为一个主要目标。大多数评论总结工作都侧重于将突出的评论作为摘要显示,这可能会忽略摘要的多样性。在本文中,我们提出了一种基于图的算法,该算法能够生成从情感角度多样化且主题覆盖良好的摘录摘要。首先,我们使用统计方法来寻找热门词汇。然后根据评论的情感类对数据集进行拆分,并对每个情感图进行排序。当与不同的基线进行比较时,我们的方法在大多数ROUGE指标中得分最高。具体来说,我们的方法显示,与最佳竞争基线相比,ROUGE-1和ROUGE-L的改进分别为3.9%和1.8%。
{"title":"Sentiment diversification for short review summarization","authors":"Mohammed Al-Dhelaan, A. Al-Suhaim","doi":"10.1145/3106426.3106525","DOIUrl":"https://doi.org/10.1145/3106426.3106525","url":null,"abstract":"With the abundance of reviews published on the Web about a given product, consumers are looking for ways to view major opinions that can be presented in a quick and succinct way. Reviews contain many different opinions, making the ability to show a diversified review summary that focus on coverage and diversity a major goal. Most review summarization work focuses on showing salient reviews as a summary which might ignore diversity in summaries. In this paper, we present a graph-based algorithm that is capable of producing extractive summaries that are both diversified from a sentiment point of view and topically well-covered. First, we use statistical measures to find topical words. Then we split the dataset based on the sentiment class of the reviews and perform the ranking on each sentiment graph. When compared with different baselines, our approach scores best in most ROUGE metrics. Specifically, our approach shows improvements of 3.9% in ROUGE-1 and 1.8% in ROUGE-L in comparison with the best competing baseline.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81614606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A logic for reasoning about evidence and belief 对证据和信念进行推理的逻辑
T. Fan, C. Liau
In agent-based systems, an agent generally forms her belief based on evidence from multiple sources, such as messages from other agents or perception of the external environment. In this paper, we present a logic for reasoning about evidence and belief. Our framework not only takes advantage of the source-tracking capability of justification logic, but also allows the distinction between the actual observation and simply potential admissibility of evidence. We present the axiomatization for the basic logic and its dynamic extension, investigate its properties, and use a running example to show its applicability to information fusion for autonomous agents.
在基于智能体的系统中,智能体通常根据来自多个来源的证据形成她的信念,例如来自其他智能体的信息或对外部环境的感知。在本文中,我们提出了一个关于证据和信念的推理逻辑。我们的框架不仅利用了证明逻辑的溯源能力,而且还允许区分实际观察和简单的潜在证据可采性。给出了基本逻辑的公理化及其动态扩展,研究了基本逻辑的性质,并用实例说明了其在自主智能体信息融合中的适用性。
{"title":"A logic for reasoning about evidence and belief","authors":"T. Fan, C. Liau","doi":"10.1145/3106426.3106519","DOIUrl":"https://doi.org/10.1145/3106426.3106519","url":null,"abstract":"In agent-based systems, an agent generally forms her belief based on evidence from multiple sources, such as messages from other agents or perception of the external environment. In this paper, we present a logic for reasoning about evidence and belief. Our framework not only takes advantage of the source-tracking capability of justification logic, but also allows the distinction between the actual observation and simply potential admissibility of evidence. We present the axiomatization for the basic logic and its dynamic extension, investigate its properties, and use a running example to show its applicability to information fusion for autonomous agents.","PeriodicalId":20685,"journal":{"name":"Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83129688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1