Yuichi Ishida, Yuma Ichikawa, Aki Dote, Toshiyuki Miyazawa, Koji Hukushima
{"title":"Ratio Divergence Learning Using Target Energy in Restricted Boltzmann Machines: Beyond Kullback--Leibler Divergence Learning","authors":"Yuichi Ishida, Yuma Ichikawa, Aki Dote, Toshiyuki Miyazawa, Koji Hukushima","doi":"arxiv-2409.07679","DOIUrl":null,"url":null,"abstract":"We propose ratio divergence (RD) learning for discrete energy-based models, a\nmethod that utilizes both training data and a tractable target energy function.\nWe apply RD learning to restricted Boltzmann machines (RBMs), which are a\nminimal model that satisfies the universal approximation theorem for discrete\ndistributions. RD learning combines the strength of both forward and reverse\nKullback-Leibler divergence (KLD) learning, effectively addressing the\n\"notorious\" issues of underfitting with the forward KLD and mode-collapse with\nthe reverse KLD. Since the summation of forward and reverse KLD seems to be\nsufficient to combine the strength of both approaches, we include this learning\nmethod as a direct baseline in numerical experiments to evaluate its\neffectiveness. Numerical experiments demonstrate that RD learning significantly\noutperforms other learning methods in terms of energy function fitting,\nmode-covering, and learning stability across various discrete energy-based\nmodels. Moreover, the performance gaps between RD learning and the other\nlearning methods become more pronounced as the dimensions of target models\nincrease.","PeriodicalId":501379,"journal":{"name":"arXiv - STAT - Statistics Theory","volume":"9 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Statistics Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07679","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We propose ratio divergence (RD) learning for discrete energy-based models, a
method that utilizes both training data and a tractable target energy function.
We apply RD learning to restricted Boltzmann machines (RBMs), which are a
minimal model that satisfies the universal approximation theorem for discrete
distributions. RD learning combines the strength of both forward and reverse
Kullback-Leibler divergence (KLD) learning, effectively addressing the
"notorious" issues of underfitting with the forward KLD and mode-collapse with
the reverse KLD. Since the summation of forward and reverse KLD seems to be
sufficient to combine the strength of both approaches, we include this learning
method as a direct baseline in numerical experiments to evaluate its
effectiveness. Numerical experiments demonstrate that RD learning significantly
outperforms other learning methods in terms of energy function fitting,
mode-covering, and learning stability across various discrete energy-based
models. Moreover, the performance gaps between RD learning and the other
learning methods become more pronounced as the dimensions of target models
increase.