A method for quantifying the generalization capabilities of generative models for solving Ising models

Qunlong Ma, Zhi Ma, Ming Gao
{"title":"A method for quantifying the generalization capabilities of generative models for solving Ising models","authors":"Qunlong Ma, Zhi Ma, Ming Gao","doi":"arxiv-2405.03435","DOIUrl":null,"url":null,"abstract":"For Ising models with complex energy landscapes, whether the ground state can\nbe found by neural networks depends heavily on the Hamming distance between the\ntraining datasets and the ground state. Despite the fact that various recently\nproposed generative models have shown good performance in solving Ising models,\nthere is no adequate discussion on how to quantify their generalization\ncapabilities. Here we design a Hamming distance regularizer in the framework of\na class of generative models, variational autoregressive networks (VAN), to\nquantify the generalization capabilities of various network architectures\ncombined with VAN. The regularizer can control the size of the overlaps between\nthe ground state and the training datasets generated by networks, which,\ntogether with the success rates of finding the ground state, form a\nquantitative metric to quantify their generalization capabilities. We conduct\nnumerical experiments on several prototypical network architectures combined\nwith VAN, including feed-forward neural networks, recurrent neural networks,\nand graph neural networks, to quantify their generalization capabilities when\nsolving Ising models. Moreover, considering the fact that the quantification of\nthe generalization capabilities of networks on small-scale problems can be used\nto predict their relative performance on large-scale problems, our method is of\ngreat significance for assisting in the Neural Architecture Search field of\nsearching for the optimal network architectures when solving large-scale Ising\nmodels.","PeriodicalId":501066,"journal":{"name":"arXiv - PHYS - Disordered Systems and Neural Networks","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - PHYS - Disordered Systems and Neural Networks","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.03435","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

For Ising models with complex energy landscapes, whether the ground state can be found by neural networks depends heavily on the Hamming distance between the training datasets and the ground state. Despite the fact that various recently proposed generative models have shown good performance in solving Ising models, there is no adequate discussion on how to quantify their generalization capabilities. Here we design a Hamming distance regularizer in the framework of a class of generative models, variational autoregressive networks (VAN), to quantify the generalization capabilities of various network architectures combined with VAN. The regularizer can control the size of the overlaps between the ground state and the training datasets generated by networks, which, together with the success rates of finding the ground state, form a quantitative metric to quantify their generalization capabilities. We conduct numerical experiments on several prototypical network architectures combined with VAN, including feed-forward neural networks, recurrent neural networks, and graph neural networks, to quantify their generalization capabilities when solving Ising models. Moreover, considering the fact that the quantification of the generalization capabilities of networks on small-scale problems can be used to predict their relative performance on large-scale problems, our method is of great significance for assisting in the Neural Architecture Search field of searching for the optimal network architectures when solving large-scale Ising models.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
量化求解伊辛模型的生成模型泛化能力的方法
对于具有复杂能谱的伊辛模型,神经网络能否找到基态在很大程度上取决于训练数据集与基态之间的汉明距离。尽管最近提出的各种生成模型在求解伊辛模型时表现出了良好的性能,但如何量化它们的泛化能力还没有充分的讨论。在此,我们在一类生成模型--变异自回归网络(VAN)的框架内设计了一个汉明距离正则器,以量化与 VAN 结合的各种网络架构的泛化能力。正则化器可以控制地面状态与网络生成的训练数据集之间的重叠大小,这与找到地面状态的成功率一起构成了量化网络泛化能力的量化指标。我们对几种与 VAN 结合的原型网络架构(包括前馈神经网络、递归神经网络和图神经网络)进行了数值实验,以量化它们在求解伊辛模型时的泛化能力。此外,考虑到量化网络在小规模问题上的泛化能力可以用来预测它们在大规模问题上的相对性能,我们的方法对于帮助神经架构搜索领域在求解大规模 Ising 模型时寻找最优网络架构具有重要意义。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Fast Analysis of the OpenAI O1-Preview Model in Solving Random K-SAT Problem: Does the LLM Solve the Problem Itself or Call an External SAT Solver? Trade-off relations between quantum coherence and measure of many-body localization Soft modes in vector spin glass models on sparse random graphs Boolean mean field spin glass model: rigorous results Generalized hetero-associative neural networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1