Monica Welfert;Gowtham R. Kurri;Kyle Otstot;Lalitha Sankar
{"title":"Addressing GAN Training Instabilities via Tunable Classification Losses","authors":"Monica Welfert;Gowtham R. Kurri;Kyle Otstot;Lalitha Sankar","doi":"10.1109/JSAIT.2024.3415670","DOIUrl":null,"url":null,"abstract":"Generative adversarial networks (GANs), modeled as a zero-sum game between a generator (G) and a discriminator (D), allow generating synthetic data with formal guarantees. Noting that D is a classifier, we begin by reformulating the GAN value function using class probability estimation (CPE) losses. We prove a two-way correspondence between CPE loss GANs and f-GANs which minimize f-divergences. We also show that all symmetric f-divergences are equivalent in convergence. In the finite sample and model capacity setting, we define and obtain bounds on estimation and generalization errors. We specialize these results to \n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n-GANs, defined using \n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n-loss, a tunable CPE loss family parametrized by \n<inline-formula> <tex-math>$\\alpha \\in (0,\\infty $ </tex-math></inline-formula>\n]. We next introduce a class of dual-objective GANs to address training instabilities of GANs by modeling each player’s objective using \n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n-loss to obtain \n<inline-formula> <tex-math>$(\\alpha _{D},\\alpha _{G})$ </tex-math></inline-formula>\n-GANs. We show that the resulting non-zero sum game simplifies to minimizing an f-divergence under appropriate conditions on \n<inline-formula> <tex-math>$(\\alpha _{D},\\alpha _{G})$ </tex-math></inline-formula>\n. Generalizing this dual-objective formulation using CPE losses, we define and obtain upper bounds on an appropriately defined estimation error. Finally, we highlight the value of tuning \n<inline-formula> <tex-math>$(\\alpha _{D},\\alpha _{G})$ </tex-math></inline-formula>\n in alleviating training instabilities for the synthetic 2D Gaussian mixture ring as well as the large publicly available Celeb-A and LSUN Classroom image datasets.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"534-553"},"PeriodicalIF":0.0000,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE journal on selected areas in information theory","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10565846/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Generative adversarial networks (GANs), modeled as a zero-sum game between a generator (G) and a discriminator (D), allow generating synthetic data with formal guarantees. Noting that D is a classifier, we begin by reformulating the GAN value function using class probability estimation (CPE) losses. We prove a two-way correspondence between CPE loss GANs and f-GANs which minimize f-divergences. We also show that all symmetric f-divergences are equivalent in convergence. In the finite sample and model capacity setting, we define and obtain bounds on estimation and generalization errors. We specialize these results to
$\alpha $
-GANs, defined using
$\alpha $
-loss, a tunable CPE loss family parametrized by
$\alpha \in (0,\infty $
]. We next introduce a class of dual-objective GANs to address training instabilities of GANs by modeling each player’s objective using
$\alpha $
-loss to obtain
$(\alpha _{D},\alpha _{G})$
-GANs. We show that the resulting non-zero sum game simplifies to minimizing an f-divergence under appropriate conditions on
$(\alpha _{D},\alpha _{G})$
. Generalizing this dual-objective formulation using CPE losses, we define and obtain upper bounds on an appropriately defined estimation error. Finally, we highlight the value of tuning
$(\alpha _{D},\alpha _{G})$
in alleviating training instabilities for the synthetic 2D Gaussian mixture ring as well as the large publicly available Celeb-A and LSUN Classroom image datasets.