Huan Fu, Mingming Gong, Chaohui Wang, Kayhan Batmanghelich, Kun Zhang, Dacheng Tao
{"title":"用于单面无监督领域映射的几何一致生成对抗网络","authors":"Huan Fu, Mingming Gong, Chaohui Wang, Kayhan Batmanghelich, Kun Zhang, Dacheng Tao","doi":"10.1109/cvpr.2019.00253","DOIUrl":null,"url":null,"abstract":"<p><p>Unsupervised domain mapping aims to learn a function G<sub>XY</sub> to translate domain <math><mi>X</mi></math> to <math><mi>Y</mi></math> in the absence of paired examples. Finding the optimal <i>G</i> <sub><i>XY</i></sub> without paired data is an ill-posed problem, so appropriate constraints are required to obtain reasonable solutions. While some prominent constraints such as cycle consistency and distance preservation successfully constrain the solution space, they overlook the special properties of images that simple geometric transformations do not change the image's semantic structure. Based on this special property, we develop a geometry-consistent generative adversarial network (<i>Gc-GAN</i>), which enables one-sided unsupervised domain mapping. <i>GcGAN</i> takes the original image and its counterpart image transformed by a predefined geometric transformation as inputs and generates two images in the new domain coupled with the corresponding geometry-consistency constraint. The geometry-consistency constraint reduces the space of possible solutions while keep the correct solutions in the search space. Quantitative and qualitative comparisons with the baseline (<i>GAN alone</i>) and the state-of-the-art methods including <i>CycleGAN</i> [66] and <i>DistanceGAN</i> [5] demonstrate the effectiveness of our method.</p>","PeriodicalId":74560,"journal":{"name":"Proceedings. IEEE Computer Society Conference on Computer Vision and Pattern Recognition","volume":"2019 ","pages":"2422-2431"},"PeriodicalIF":0.0000,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7030214/pdf/nihms-1037392.pdf","citationCount":"0","resultStr":"{\"title\":\"Geometry-Consistent Generative Adversarial Networks for One-Sided Unsupervised Domain Mapping.\",\"authors\":\"Huan Fu, Mingming Gong, Chaohui Wang, Kayhan Batmanghelich, Kun Zhang, Dacheng Tao\",\"doi\":\"10.1109/cvpr.2019.00253\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Unsupervised domain mapping aims to learn a function G<sub>XY</sub> to translate domain <math><mi>X</mi></math> to <math><mi>Y</mi></math> in the absence of paired examples. Finding the optimal <i>G</i> <sub><i>XY</i></sub> without paired data is an ill-posed problem, so appropriate constraints are required to obtain reasonable solutions. While some prominent constraints such as cycle consistency and distance preservation successfully constrain the solution space, they overlook the special properties of images that simple geometric transformations do not change the image's semantic structure. Based on this special property, we develop a geometry-consistent generative adversarial network (<i>Gc-GAN</i>), which enables one-sided unsupervised domain mapping. <i>GcGAN</i> takes the original image and its counterpart image transformed by a predefined geometric transformation as inputs and generates two images in the new domain coupled with the corresponding geometry-consistency constraint. The geometry-consistency constraint reduces the space of possible solutions while keep the correct solutions in the search space. Quantitative and qualitative comparisons with the baseline (<i>GAN alone</i>) and the state-of-the-art methods including <i>CycleGAN</i> [66] and <i>DistanceGAN</i> [5] demonstrate the effectiveness of our method.</p>\",\"PeriodicalId\":74560,\"journal\":{\"name\":\"Proceedings. IEEE Computer Society Conference on Computer Vision and Pattern Recognition\",\"volume\":\"2019 \",\"pages\":\"2422-2431\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7030214/pdf/nihms-1037392.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings. IEEE Computer Society Conference on Computer Vision and Pattern Recognition\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/cvpr.2019.00253\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2020/1/9 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings. IEEE Computer Society Conference on Computer Vision and Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/cvpr.2019.00253","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2020/1/9 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
无监督领域映射旨在学习一个函数 GXY,以便在没有配对示例的情况下将领域 X 转换为 Y。在没有配对数据的情况下寻找最优 G XY 是一个难以解决的问题,因此需要适当的约束条件才能获得合理的解决方案。虽然一些著名的约束条件(如周期一致性和距离保持)成功地限制了解空间,但它们忽略了图像的特殊属性,即简单的几何变换不会改变图像的语义结构。基于这一特殊属性,我们开发了一种几何一致性生成对抗网络(Gc-GAN),它可以实现单侧无监督领域映射。GcGAN 将原始图像和经过预定义几何变换的对应图像作为输入,并在新域中生成两幅图像以及相应的几何一致性约束。几何一致性约束减少了可能解决方案的空间,同时在搜索空间中保留了正确的解决方案。与基线(单独的 GAN)和最先进的方法(包括 CycleGAN [66] 和 DistanceGAN [5])进行的定量和定性比较证明了我们方法的有效性。
Geometry-Consistent Generative Adversarial Networks for One-Sided Unsupervised Domain Mapping.
Unsupervised domain mapping aims to learn a function GXY to translate domain to in the absence of paired examples. Finding the optimal GXY without paired data is an ill-posed problem, so appropriate constraints are required to obtain reasonable solutions. While some prominent constraints such as cycle consistency and distance preservation successfully constrain the solution space, they overlook the special properties of images that simple geometric transformations do not change the image's semantic structure. Based on this special property, we develop a geometry-consistent generative adversarial network (Gc-GAN), which enables one-sided unsupervised domain mapping. GcGAN takes the original image and its counterpart image transformed by a predefined geometric transformation as inputs and generates two images in the new domain coupled with the corresponding geometry-consistency constraint. The geometry-consistency constraint reduces the space of possible solutions while keep the correct solutions in the search space. Quantitative and qualitative comparisons with the baseline (GAN alone) and the state-of-the-art methods including CycleGAN [66] and DistanceGAN [5] demonstrate the effectiveness of our method.