N. Xu, Kangkang Song, Jiangjian Xiao, Chengbin Peng
{"title":"图像翻译中保持视觉感知的去噪网络","authors":"N. Xu, Kangkang Song, Jiangjian Xiao, Chengbin Peng","doi":"10.1109/ISPDS56360.2022.9874112","DOIUrl":null,"url":null,"abstract":"Image denoising is a fundamental problem in computer vision and has received much attention from scholars. With the fast development of convolutional neural networks, more and more deep learning-based noise reduction algorithms have emerged. However, current image denoising networks tend to apply image noise reduction only in the RGB color space, ignoring the information at the visual perception level, making the images generated by these algorithms too smooth and lacking texture and details. Therefore, this paper proposes a novel noise reduction network in the image translation area using deep learning feature space instead of the traditional RGB color space to restore more realistic and more detailed texture information in generated images. The network contains a visual perception generator and a multi-objective optimization network. The generator includes a multiscale encoding-decoding sub-network, which extracts high-level perception features from input images. The optimization network contains content consistency loss, multiscale adversarial generation loss, and discriminator feature alignment loss, which effectively retains detailed texture information in the images. We synthesized noise of suitable intensity based on publicly available data sets and conducted multiple experiments to verify the effectiveness of the algorithm. The experimental results show that the proposed algorithm significantly improves textures and details in denoised images. The algorithm removes a large amount of noise information while preserving lots of perceptual information at the visual level, generating more realistic images with detailed texture features.","PeriodicalId":280244,"journal":{"name":"2022 3rd International Conference on Information Science, Parallel and Distributed Systems (ISPDS)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"visual perception preserved denoising network in Image translation\",\"authors\":\"N. Xu, Kangkang Song, Jiangjian Xiao, Chengbin Peng\",\"doi\":\"10.1109/ISPDS56360.2022.9874112\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Image denoising is a fundamental problem in computer vision and has received much attention from scholars. With the fast development of convolutional neural networks, more and more deep learning-based noise reduction algorithms have emerged. However, current image denoising networks tend to apply image noise reduction only in the RGB color space, ignoring the information at the visual perception level, making the images generated by these algorithms too smooth and lacking texture and details. Therefore, this paper proposes a novel noise reduction network in the image translation area using deep learning feature space instead of the traditional RGB color space to restore more realistic and more detailed texture information in generated images. The network contains a visual perception generator and a multi-objective optimization network. The generator includes a multiscale encoding-decoding sub-network, which extracts high-level perception features from input images. The optimization network contains content consistency loss, multiscale adversarial generation loss, and discriminator feature alignment loss, which effectively retains detailed texture information in the images. We synthesized noise of suitable intensity based on publicly available data sets and conducted multiple experiments to verify the effectiveness of the algorithm. The experimental results show that the proposed algorithm significantly improves textures and details in denoised images. The algorithm removes a large amount of noise information while preserving lots of perceptual information at the visual level, generating more realistic images with detailed texture features.\",\"PeriodicalId\":280244,\"journal\":{\"name\":\"2022 3rd International Conference on Information Science, Parallel and Distributed Systems (ISPDS)\",\"volume\":\"33 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-07-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 3rd International Conference on Information Science, Parallel and Distributed Systems (ISPDS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISPDS56360.2022.9874112\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 3rd International Conference on Information Science, Parallel and Distributed Systems (ISPDS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISPDS56360.2022.9874112","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
visual perception preserved denoising network in Image translation
Image denoising is a fundamental problem in computer vision and has received much attention from scholars. With the fast development of convolutional neural networks, more and more deep learning-based noise reduction algorithms have emerged. However, current image denoising networks tend to apply image noise reduction only in the RGB color space, ignoring the information at the visual perception level, making the images generated by these algorithms too smooth and lacking texture and details. Therefore, this paper proposes a novel noise reduction network in the image translation area using deep learning feature space instead of the traditional RGB color space to restore more realistic and more detailed texture information in generated images. The network contains a visual perception generator and a multi-objective optimization network. The generator includes a multiscale encoding-decoding sub-network, which extracts high-level perception features from input images. The optimization network contains content consistency loss, multiscale adversarial generation loss, and discriminator feature alignment loss, which effectively retains detailed texture information in the images. We synthesized noise of suitable intensity based on publicly available data sets and conducted multiple experiments to verify the effectiveness of the algorithm. The experimental results show that the proposed algorithm significantly improves textures and details in denoised images. The algorithm removes a large amount of noise information while preserving lots of perceptual information at the visual level, generating more realistic images with detailed texture features.