{"title":"基于多鉴别器生成对抗网络的图像自动着色","authors":"Youssef Mourchid, M. Donias, Y. Berthoumieu","doi":"10.23919/Eusipco47968.2020.9287792","DOIUrl":null,"url":null,"abstract":"This paper presents a deep automatic colorization approach which avoids any manual intervention. Recently Generative Adversarial Network (GANs) approaches have proven their effectiveness for image colorization tasks. Inspired by GANs methods, we propose a novel colorization model that produces more realistic quality results. The model employs an additional discriminator which works in the feature domain. Using a feature discriminator, our generator produces structural high-frequency features instead of noisy artifacts. To achieve the required level of details in the colorization process, we incorporate non-adversarial losses from recent image style transfer techniques. Besides, the generator architecture follows the general shape of U-Net, to transfer information more effectively between distant layers. The performance of the proposed model was evaluated quantitatively as well as qualitatively with places365 dataset. Results show that the proposed model achieves more realistic colors with less artifacts compared to the state-of-the-art approaches.","PeriodicalId":6705,"journal":{"name":"2020 28th European Signal Processing Conference (EUSIPCO)","volume":"23 1","pages":"1532-1536"},"PeriodicalIF":0.0000,"publicationDate":"2021-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Automatic Image Colorization based on Multi-Discriminators Generative Adversarial Networks\",\"authors\":\"Youssef Mourchid, M. Donias, Y. Berthoumieu\",\"doi\":\"10.23919/Eusipco47968.2020.9287792\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper presents a deep automatic colorization approach which avoids any manual intervention. Recently Generative Adversarial Network (GANs) approaches have proven their effectiveness for image colorization tasks. Inspired by GANs methods, we propose a novel colorization model that produces more realistic quality results. The model employs an additional discriminator which works in the feature domain. Using a feature discriminator, our generator produces structural high-frequency features instead of noisy artifacts. To achieve the required level of details in the colorization process, we incorporate non-adversarial losses from recent image style transfer techniques. Besides, the generator architecture follows the general shape of U-Net, to transfer information more effectively between distant layers. The performance of the proposed model was evaluated quantitatively as well as qualitatively with places365 dataset. Results show that the proposed model achieves more realistic colors with less artifacts compared to the state-of-the-art approaches.\",\"PeriodicalId\":6705,\"journal\":{\"name\":\"2020 28th European Signal Processing Conference (EUSIPCO)\",\"volume\":\"23 1\",\"pages\":\"1532-1536\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-01-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 28th European Signal Processing Conference (EUSIPCO)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.23919/Eusipco47968.2020.9287792\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 28th European Signal Processing Conference (EUSIPCO)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/Eusipco47968.2020.9287792","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Automatic Image Colorization based on Multi-Discriminators Generative Adversarial Networks
This paper presents a deep automatic colorization approach which avoids any manual intervention. Recently Generative Adversarial Network (GANs) approaches have proven their effectiveness for image colorization tasks. Inspired by GANs methods, we propose a novel colorization model that produces more realistic quality results. The model employs an additional discriminator which works in the feature domain. Using a feature discriminator, our generator produces structural high-frequency features instead of noisy artifacts. To achieve the required level of details in the colorization process, we incorporate non-adversarial losses from recent image style transfer techniques. Besides, the generator architecture follows the general shape of U-Net, to transfer information more effectively between distant layers. The performance of the proposed model was evaluated quantitatively as well as qualitatively with places365 dataset. Results show that the proposed model achieves more realistic colors with less artifacts compared to the state-of-the-art approaches.