{"title":"Infrared Image Generation Algorithm Based on GAN and contrastive learning","authors":"Hong Liu, Lei Ma","doi":"10.1109/AICIT55386.2022.9930233","DOIUrl":null,"url":null,"abstract":"For the task of converting dimly lit, low luminance nighttime visible to infrared images, we propose a Contrastive Visible-Infrared Image Translation Network (CVIIT). To better distinguish and translate objects such as pedestrians and vehicles, we introduce an attention module based on class activation map in the generator and discriminator of the Generative Adversarial Network (GAN), which captures richer context information in the images. In addition, we introduce contrastive learning to align the generated images with the visible images in terms of content. Qualitative and quantitative experiments on a publicly available visible-Infrared image pairing dataset (LLVIP) show that the proposed method generates infrared images of significantly higher quality than other state-of-the-art image-to-image translation (I2IT) methods.","PeriodicalId":231070,"journal":{"name":"2022 International Conference on Artificial Intelligence and Computer Information Technology (AICIT)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Artificial Intelligence and Computer Information Technology (AICIT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AICIT55386.2022.9930233","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
For the task of converting dimly lit, low luminance nighttime visible to infrared images, we propose a Contrastive Visible-Infrared Image Translation Network (CVIIT). To better distinguish and translate objects such as pedestrians and vehicles, we introduce an attention module based on class activation map in the generator and discriminator of the Generative Adversarial Network (GAN), which captures richer context information in the images. In addition, we introduce contrastive learning to align the generated images with the visible images in terms of content. Qualitative and quantitative experiments on a publicly available visible-Infrared image pairing dataset (LLVIP) show that the proposed method generates infrared images of significantly higher quality than other state-of-the-art image-to-image translation (I2IT) methods.