Knowledge Embedding towards the Recommendation with Sparse User-Item Interactions

Deqing Yang, Ziyi Wang, Junyan Jiang, Yanghua Xiao
{"title":"Knowledge Embedding towards the Recommendation with Sparse User-Item Interactions","authors":"Deqing Yang, Ziyi Wang, Junyan Jiang, Yanghua Xiao","doi":"10.1145/3341161.3342876","DOIUrl":null,"url":null,"abstract":"Recently, many researchers in recommender systems have realized that encoding user-item interactions based on deep neural networks (DNNs) promotes collaborative-filtering (CF)'s performance. Nonetheless, those DNN-based models' performance is still limited when observed user-item interactions are very less because the training samples distilled from these interactions are critical for deep learning models. To address this problem, we resort to plenty features distilled from knowledge graphs (KGs), to profile users and items precisely and sufficiently rather than observed user-item interactions. In this paper, we propose a knowledge embedding based recommendation framework to alleviate the problem of sparse user-item interactions in recommendation. In our framework, each user and each item are both represented by the combination of an item embedding and a tag embedding at first. Specifically, item embeddings are learned by Metapath2Vec which is a graph embedding model qualified to embedding heterogeneous information networks. Tag embeddings are learned by a Skip-gram model similar to word embedding. We regarded these embeddings as knowledge embeddings because they both indicate knowledge about the latent relationships of movie-movie and user-movie. At last, a target user's representation and a candidate movie's representation are both fed into a multi-layer perceptron to output the probability that the user likes the item. The probability can be further used to achieve top-n recommendation. The extensive experiments on a movie recommendation dataset demonstrate our framework's superiority over some state-of-the-art recommendation models, especially in the scenario of sparse user-movie interactions.","PeriodicalId":403360,"journal":{"name":"2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM)","volume":"136 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3341161.3342876","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 11

Abstract

Recently, many researchers in recommender systems have realized that encoding user-item interactions based on deep neural networks (DNNs) promotes collaborative-filtering (CF)'s performance. Nonetheless, those DNN-based models' performance is still limited when observed user-item interactions are very less because the training samples distilled from these interactions are critical for deep learning models. To address this problem, we resort to plenty features distilled from knowledge graphs (KGs), to profile users and items precisely and sufficiently rather than observed user-item interactions. In this paper, we propose a knowledge embedding based recommendation framework to alleviate the problem of sparse user-item interactions in recommendation. In our framework, each user and each item are both represented by the combination of an item embedding and a tag embedding at first. Specifically, item embeddings are learned by Metapath2Vec which is a graph embedding model qualified to embedding heterogeneous information networks. Tag embeddings are learned by a Skip-gram model similar to word embedding. We regarded these embeddings as knowledge embeddings because they both indicate knowledge about the latent relationships of movie-movie and user-movie. At last, a target user's representation and a candidate movie's representation are both fed into a multi-layer perceptron to output the probability that the user likes the item. The probability can be further used to achieve top-n recommendation. The extensive experiments on a movie recommendation dataset demonstrate our framework's superiority over some state-of-the-art recommendation models, especially in the scenario of sparse user-movie interactions.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
面向稀疏用户-项目交互推荐的知识嵌入
近年来,许多推荐系统的研究人员已经意识到,基于深度神经网络(dnn)对用户-物品交互进行编码可以提高协同过滤(CF)的性能。尽管如此,当观察到的用户-物品交互非常少时,这些基于dnn的模型的性能仍然有限,因为从这些交互中提取的训练样本对深度学习模型至关重要。为了解决这个问题,我们从知识图(KGs)中提取了大量的特征,以准确而充分地描述用户和项目,而不是观察用户和项目的交互。本文提出了一种基于知识嵌入的推荐框架,以缓解推荐中用户-项目交互稀疏的问题。在我们的框架中,每个用户和每个项目首先都由项目嵌入和标签嵌入的组合来表示。具体来说,项目嵌入是通过Metapath2Vec学习的,Metapath2Vec是一种适合嵌入异构信息网络的图嵌入模型。标签嵌入是通过类似于词嵌入的Skip-gram模型来学习的。我们将这些嵌入视为知识嵌入,因为它们都表示关于电影-电影和用户-电影潜在关系的知识。最后,将目标用户的表示和候选电影的表示都输入到多层感知器中,以输出用户喜欢该项目的概率。该概率可以进一步用于实现top-n推荐。在电影推荐数据集上的大量实验表明,我们的框架优于一些最先进的推荐模型,特别是在稀疏的用户-电影交互场景中。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Neural-Brane: An inductive approach for attributed network embedding Customer Recommendation Based on Profile Matching and Customized Campaigns in On-Line Social Networks Characterizing and Detecting Livestreaming Chatbots Two Decades of Network Science: as seen through the co-authorship network of network scientists Show me your friends, and I will tell you whom you vote for: Predicting voting behavior in social networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1