{"title":"Daisee: Adaptive importance sampling by balancing exploration and exploitation","authors":"Xiaoyu Lu, Tom Rainforth, Y. Teh","doi":"10.1111/sjos.12637","DOIUrl":null,"url":null,"abstract":"We study adaptive importance sampling (AIS) as an online learning problem and argue for the importance of the trade‐off between exploration and exploitation in this adaptation. Borrowing ideas from the online learning literature, we propose Daisee, a partition‐based AIS algorithm. We further introduce a notion of regret for AIS and show that Daisee has 𝒪(T(logT)34) cumulative pseudo‐regret, where T$$ T $$ is the number of iterations. We then extend Daisee to adaptively learn a hierarchical partitioning of the sample space for more efficient sampling and confirm the performance of both algorithms empirically.","PeriodicalId":0,"journal":{"name":"","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1111/sjos.12637","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We study adaptive importance sampling (AIS) as an online learning problem and argue for the importance of the trade‐off between exploration and exploitation in this adaptation. Borrowing ideas from the online learning literature, we propose Daisee, a partition‐based AIS algorithm. We further introduce a notion of regret for AIS and show that Daisee has 𝒪(T(logT)34) cumulative pseudo‐regret, where T$$ T $$ is the number of iterations. We then extend Daisee to adaptively learn a hierarchical partitioning of the sample space for more efficient sampling and confirm the performance of both algorithms empirically.
我们将自适应重要性抽样(AIS)作为一个在线学习问题来研究,并论证了在这种适应中探索和利用之间权衡的重要性。借鉴在线学习文献的思想,我们提出了一种基于分割的AIS算法Daisee。我们进一步为AIS引入了后悔的概念,并表明Daisee具有 (T(logT)34)累积伪后悔,其中T $$ T $$是迭代次数。然后,我们扩展Daisee自适应学习样本空间的分层划分,以获得更有效的采样,并通过经验验证两种算法的性能。