Monica Welfert;Gowtham R. Kurri;Kyle Otstot;Lalitha Sankar
{"title":"通过可调分类损失解决 GAN 训练不稳定性问题","authors":"Monica Welfert;Gowtham R. Kurri;Kyle Otstot;Lalitha Sankar","doi":"10.1109/JSAIT.2024.3415670","DOIUrl":null,"url":null,"abstract":"Generative adversarial networks (GANs), modeled as a zero-sum game between a generator (G) and a discriminator (D), allow generating synthetic data with formal guarantees. Noting that D is a classifier, we begin by reformulating the GAN value function using class probability estimation (CPE) losses. We prove a two-way correspondence between CPE loss GANs and f-GANs which minimize f-divergences. We also show that all symmetric f-divergences are equivalent in convergence. In the finite sample and model capacity setting, we define and obtain bounds on estimation and generalization errors. We specialize these results to \n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n-GANs, defined using \n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n-loss, a tunable CPE loss family parametrized by \n<inline-formula> <tex-math>$\\alpha \\in (0,\\infty $ </tex-math></inline-formula>\n]. We next introduce a class of dual-objective GANs to address training instabilities of GANs by modeling each player’s objective using \n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n-loss to obtain \n<inline-formula> <tex-math>$(\\alpha _{D},\\alpha _{G})$ </tex-math></inline-formula>\n-GANs. We show that the resulting non-zero sum game simplifies to minimizing an f-divergence under appropriate conditions on \n<inline-formula> <tex-math>$(\\alpha _{D},\\alpha _{G})$ </tex-math></inline-formula>\n. Generalizing this dual-objective formulation using CPE losses, we define and obtain upper bounds on an appropriately defined estimation error. Finally, we highlight the value of tuning \n<inline-formula> <tex-math>$(\\alpha _{D},\\alpha _{G})$ </tex-math></inline-formula>\n in alleviating training instabilities for the synthetic 2D Gaussian mixture ring as well as the large publicly available Celeb-A and LSUN Classroom image datasets.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"534-553"},"PeriodicalIF":0.0000,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Addressing GAN Training Instabilities via Tunable Classification Losses\",\"authors\":\"Monica Welfert;Gowtham R. Kurri;Kyle Otstot;Lalitha Sankar\",\"doi\":\"10.1109/JSAIT.2024.3415670\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Generative adversarial networks (GANs), modeled as a zero-sum game between a generator (G) and a discriminator (D), allow generating synthetic data with formal guarantees. Noting that D is a classifier, we begin by reformulating the GAN value function using class probability estimation (CPE) losses. We prove a two-way correspondence between CPE loss GANs and f-GANs which minimize f-divergences. We also show that all symmetric f-divergences are equivalent in convergence. In the finite sample and model capacity setting, we define and obtain bounds on estimation and generalization errors. We specialize these results to \\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n-GANs, defined using \\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n-loss, a tunable CPE loss family parametrized by \\n<inline-formula> <tex-math>$\\\\alpha \\\\in (0,\\\\infty $ </tex-math></inline-formula>\\n]. We next introduce a class of dual-objective GANs to address training instabilities of GANs by modeling each player’s objective using \\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n-loss to obtain \\n<inline-formula> <tex-math>$(\\\\alpha _{D},\\\\alpha _{G})$ </tex-math></inline-formula>\\n-GANs. We show that the resulting non-zero sum game simplifies to minimizing an f-divergence under appropriate conditions on \\n<inline-formula> <tex-math>$(\\\\alpha _{D},\\\\alpha _{G})$ </tex-math></inline-formula>\\n. Generalizing this dual-objective formulation using CPE losses, we define and obtain upper bounds on an appropriately defined estimation error. Finally, we highlight the value of tuning \\n<inline-formula> <tex-math>$(\\\\alpha _{D},\\\\alpha _{G})$ </tex-math></inline-formula>\\n in alleviating training instabilities for the synthetic 2D Gaussian mixture ring as well as the large publicly available Celeb-A and LSUN Classroom image datasets.\",\"PeriodicalId\":73295,\"journal\":{\"name\":\"IEEE journal on selected areas in information theory\",\"volume\":\"5 \",\"pages\":\"534-553\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-06-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE journal on selected areas in information theory\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10565846/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE journal on selected areas in information theory","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10565846/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Addressing GAN Training Instabilities via Tunable Classification Losses
Generative adversarial networks (GANs), modeled as a zero-sum game between a generator (G) and a discriminator (D), allow generating synthetic data with formal guarantees. Noting that D is a classifier, we begin by reformulating the GAN value function using class probability estimation (CPE) losses. We prove a two-way correspondence between CPE loss GANs and f-GANs which minimize f-divergences. We also show that all symmetric f-divergences are equivalent in convergence. In the finite sample and model capacity setting, we define and obtain bounds on estimation and generalization errors. We specialize these results to
$\alpha $
-GANs, defined using
$\alpha $
-loss, a tunable CPE loss family parametrized by
$\alpha \in (0,\infty $
]. We next introduce a class of dual-objective GANs to address training instabilities of GANs by modeling each player’s objective using
$\alpha $
-loss to obtain
$(\alpha _{D},\alpha _{G})$
-GANs. We show that the resulting non-zero sum game simplifies to minimizing an f-divergence under appropriate conditions on
$(\alpha _{D},\alpha _{G})$
. Generalizing this dual-objective formulation using CPE losses, we define and obtain upper bounds on an appropriately defined estimation error. Finally, we highlight the value of tuning
$(\alpha _{D},\alpha _{G})$
in alleviating training instabilities for the synthetic 2D Gaussian mixture ring as well as the large publicly available Celeb-A and LSUN Classroom image datasets.