Bayesian Low-Rank Determinantal Point Processes

Mike Gartrell, U. Paquet, Noam Koenigstein
{"title":"Bayesian Low-Rank Determinantal Point Processes","authors":"Mike Gartrell, U. Paquet, Noam Koenigstein","doi":"10.1145/2959100.2959178","DOIUrl":null,"url":null,"abstract":"Determinantal point processes (DPPs) are an emerging model for encoding probabilities over subsets, such as shopping baskets, selected from a ground set, such as an item catalog. They have recently proved to be appealing models for a number of machine learning tasks, including product recommendation. DPPs are parametrized by a positive semi-definite kernel matrix. Prior work has shown that using a low-rank factorization of this kernel provides scalability improvements that open the door to training on large-scale datasets and computing online recommendations, both of which are infeasible with standard DPP models that use a full-rank kernel. A low-rank DPP model can be trained using an optimization-based method, such as stochastic gradient ascent, to find a point estimate of the kernel parameters, which can be performed efficiently on large-scale datasets. However, this approach requires careful tuning of regularization parameters to prevent overfitting and provide good predictive performance, which can be computationally expensive. In this paper we present a Bayesian method for learning a low-rank factorization of this kernel, which provides automatic control of regularization. We show that our Bayesian low-rank DPP model can be trained efficiently using stochastic gradient Hamiltonian Monte Carlo (SGHMC). Our Bayesian model generally provides better predictive performance on several real-world product recommendation datasets than optimization-based low-rank DPP models trained using stochastic gradient ascent, and better performance than several state-of-the art recommendation methods in many cases.","PeriodicalId":315651,"journal":{"name":"Proceedings of the 10th ACM Conference on Recommender Systems","volume":"47 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"59","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 10th ACM Conference on Recommender Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2959100.2959178","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 59

Abstract

Determinantal point processes (DPPs) are an emerging model for encoding probabilities over subsets, such as shopping baskets, selected from a ground set, such as an item catalog. They have recently proved to be appealing models for a number of machine learning tasks, including product recommendation. DPPs are parametrized by a positive semi-definite kernel matrix. Prior work has shown that using a low-rank factorization of this kernel provides scalability improvements that open the door to training on large-scale datasets and computing online recommendations, both of which are infeasible with standard DPP models that use a full-rank kernel. A low-rank DPP model can be trained using an optimization-based method, such as stochastic gradient ascent, to find a point estimate of the kernel parameters, which can be performed efficiently on large-scale datasets. However, this approach requires careful tuning of regularization parameters to prevent overfitting and provide good predictive performance, which can be computationally expensive. In this paper we present a Bayesian method for learning a low-rank factorization of this kernel, which provides automatic control of regularization. We show that our Bayesian low-rank DPP model can be trained efficiently using stochastic gradient Hamiltonian Monte Carlo (SGHMC). Our Bayesian model generally provides better predictive performance on several real-world product recommendation datasets than optimization-based low-rank DPP models trained using stochastic gradient ascent, and better performance than several state-of-the art recommendation methods in many cases.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
贝叶斯低秩行列式点过程
决定性点过程(dpp)是一种新兴的模型,用于对子集(如购物篮)上的概率进行编码,这些子集是从基础集(如商品目录)中选择的。它们最近被证明是许多机器学习任务的有吸引力的模型,包括产品推荐。用半正定核矩阵对dpp进行参数化。先前的工作表明,使用该内核的低秩分解提供了可扩展性的改进,为大规模数据集的训练和计算在线推荐打开了大门,这两者对于使用全秩内核的标准DPP模型都是不可行的。低秩DPP模型可以使用基于优化的方法(如随机梯度上升)进行训练,以找到核参数的点估计,这可以在大规模数据集上有效地执行。然而,这种方法需要仔细调整正则化参数,以防止过拟合并提供良好的预测性能,这在计算上可能是昂贵的。在本文中,我们提出了一种贝叶斯方法来学习这种核的低秩分解,它提供了正则化的自动控制。我们证明贝叶斯低秩DPP模型可以使用随机梯度哈密顿蒙特卡罗(SGHMC)有效地训练。我们的贝叶斯模型在一些现实世界的产品推荐数据集上通常比使用随机梯度上升训练的基于优化的低秩DPP模型提供更好的预测性能,并且在许多情况下比几种最先进的推荐方法表现更好。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Opening Remarks Mining Information for the Cold-Item Problem Are You Influenced by Others When Rating?: Improve Rating Prediction by Conformity Modeling Contrasting Offline and Online Results when Evaluating Recommendation Algorithms Intent-Aware Diversification Using a Constrained PLSA
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1