{"title":"ECycleGAN","authors":"Xianchao Zhang, Changjia Zhou","doi":"10.1145/3404555.3404597","DOIUrl":null,"url":null,"abstract":"Unsupervised image-to-image translation, which aims in translating two irrelevant domains of images, has increased substantially in recent years with the success of Generative Adversarial Networks (GANs) based on the cycle-consistency assumption. Especially, the Cycle-Consistent Generative Adversarial Network (CycleGAN) has shown remarkable success for two domains translation. However, the details about texture and style are often accompanied with unpleasant artifacts. To further enhance the translational quality, we thoroughly study the key components of CycleGAN - network architecture and adversarial loss, and improve each of them to derive an Enhanced CycleGAN (ECycleGAN). In particular, we propose a perceptual loss function which motivated by perceptual similarity instead of similarity in pixel space. Moreover, we introduce the Residual Dense Normalization Block (RDNB) to replace the residual blocks as the basic network building unit. Finally, we borrow the idea from WGAN-GP as the adversarial loss functions. The ECycleGAN, thanks to these changes, demonstrates appealing visual quality with more realistic and natural textures than any state-of-the-art methods.","PeriodicalId":220526,"journal":{"name":"Proceedings of the 2020 6th International Conference on Computing and Artificial Intelligence","volume":"63 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2020 6th International Conference on Computing and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3404555.3404597","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Unsupervised image-to-image translation, which aims in translating two irrelevant domains of images, has increased substantially in recent years with the success of Generative Adversarial Networks (GANs) based on the cycle-consistency assumption. Especially, the Cycle-Consistent Generative Adversarial Network (CycleGAN) has shown remarkable success for two domains translation. However, the details about texture and style are often accompanied with unpleasant artifacts. To further enhance the translational quality, we thoroughly study the key components of CycleGAN - network architecture and adversarial loss, and improve each of them to derive an Enhanced CycleGAN (ECycleGAN). In particular, we propose a perceptual loss function which motivated by perceptual similarity instead of similarity in pixel space. Moreover, we introduce the Residual Dense Normalization Block (RDNB) to replace the residual blocks as the basic network building unit. Finally, we borrow the idea from WGAN-GP as the adversarial loss functions. The ECycleGAN, thanks to these changes, demonstrates appealing visual quality with more realistic and natural textures than any state-of-the-art methods.