Xiaoyang Shen , Haibin Li , Yaqian Li , Wenming Zhang
{"title":"利用多域特征增强低照度图像的细节和纹理","authors":"Xiaoyang Shen , Haibin Li , Yaqian Li , Wenming Zhang","doi":"10.1016/j.dsp.2024.104808","DOIUrl":null,"url":null,"abstract":"<div><div>Low-light image enhancement holds significant value in the fields of computer vision and image processing, such as in applications like surveillance photography and medical imaging. Images captured in low-light environments typically suffer from significant noise levels, low contrast, and color distortion. Although existing low-light image enhancement techniques can improve image brightness and contrast to some extent, they often introduce noise or result in over-enhancement, leading to the loss of detail and texture. This paper introduces an innovative approach to low-light image enhancement by fusing spatial and frequency domain features while optimizing them with multiple loss functions. The core of the algorithm lies in its multi-branch feature extraction, multi-loss function constraints, and a carefully designed model structure. In particular, the model employs an encoder-decoder architecture, where the encoder extracts spatial features from the image, the Fourier feature extraction module captures frequency domain information, and the histogram feature encoder-decoder module processes global brightness distribution. These extracted features are then fused and reconstructed in the decoder to produce the enhanced image. In terms of loss functions, the algorithm combines perceptual loss, structural similarity loss, Fourier loss, and histogram loss to ensure comprehensive and natural enhancement effects. The novelty of this algorithm lies not only in its multi-branch feature extraction design but also in its unique model structure, which synergistically improves image quality across different domains, effectively preventing over-enhancement, and ultimately achieving a balanced enhancement of brightness, details, and texture. Experimental results on multiple datasets, including SIDD, LOL, MIT-Adobe-FiveK, and LOL-v2-synthetic, demonstrate that the proposed method outperforms existing techniques in terms of image detail, texture, and brightness. Specifically, it achieves a PSNR of 27.52 dB on the LOL dataset, surpassing Wavelet Diffusion by 1.19 dB. Additionally, on the LOL-v2-synthetic dataset, it achieves a PSNR of 29.56 dB, exceeding Wavelet Diffusion by 3.06 dB. These results demonstrate a significant enhancement in the visual quality of low-light images.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"156 ","pages":"Article 104808"},"PeriodicalIF":2.9000,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Low-light image enhancement guided by multi-domain features for detail and texture enhancement\",\"authors\":\"Xiaoyang Shen , Haibin Li , Yaqian Li , Wenming Zhang\",\"doi\":\"10.1016/j.dsp.2024.104808\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Low-light image enhancement holds significant value in the fields of computer vision and image processing, such as in applications like surveillance photography and medical imaging. Images captured in low-light environments typically suffer from significant noise levels, low contrast, and color distortion. Although existing low-light image enhancement techniques can improve image brightness and contrast to some extent, they often introduce noise or result in over-enhancement, leading to the loss of detail and texture. This paper introduces an innovative approach to low-light image enhancement by fusing spatial and frequency domain features while optimizing them with multiple loss functions. The core of the algorithm lies in its multi-branch feature extraction, multi-loss function constraints, and a carefully designed model structure. In particular, the model employs an encoder-decoder architecture, where the encoder extracts spatial features from the image, the Fourier feature extraction module captures frequency domain information, and the histogram feature encoder-decoder module processes global brightness distribution. These extracted features are then fused and reconstructed in the decoder to produce the enhanced image. In terms of loss functions, the algorithm combines perceptual loss, structural similarity loss, Fourier loss, and histogram loss to ensure comprehensive and natural enhancement effects. The novelty of this algorithm lies not only in its multi-branch feature extraction design but also in its unique model structure, which synergistically improves image quality across different domains, effectively preventing over-enhancement, and ultimately achieving a balanced enhancement of brightness, details, and texture. Experimental results on multiple datasets, including SIDD, LOL, MIT-Adobe-FiveK, and LOL-v2-synthetic, demonstrate that the proposed method outperforms existing techniques in terms of image detail, texture, and brightness. Specifically, it achieves a PSNR of 27.52 dB on the LOL dataset, surpassing Wavelet Diffusion by 1.19 dB. Additionally, on the LOL-v2-synthetic dataset, it achieves a PSNR of 29.56 dB, exceeding Wavelet Diffusion by 3.06 dB. These results demonstrate a significant enhancement in the visual quality of low-light images.</div></div>\",\"PeriodicalId\":51011,\"journal\":{\"name\":\"Digital Signal Processing\",\"volume\":\"156 \",\"pages\":\"Article 104808\"},\"PeriodicalIF\":2.9000,\"publicationDate\":\"2024-10-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Digital Signal Processing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1051200424004330\",\"RegionNum\":3,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Digital Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1051200424004330","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Low-light image enhancement guided by multi-domain features for detail and texture enhancement
Low-light image enhancement holds significant value in the fields of computer vision and image processing, such as in applications like surveillance photography and medical imaging. Images captured in low-light environments typically suffer from significant noise levels, low contrast, and color distortion. Although existing low-light image enhancement techniques can improve image brightness and contrast to some extent, they often introduce noise or result in over-enhancement, leading to the loss of detail and texture. This paper introduces an innovative approach to low-light image enhancement by fusing spatial and frequency domain features while optimizing them with multiple loss functions. The core of the algorithm lies in its multi-branch feature extraction, multi-loss function constraints, and a carefully designed model structure. In particular, the model employs an encoder-decoder architecture, where the encoder extracts spatial features from the image, the Fourier feature extraction module captures frequency domain information, and the histogram feature encoder-decoder module processes global brightness distribution. These extracted features are then fused and reconstructed in the decoder to produce the enhanced image. In terms of loss functions, the algorithm combines perceptual loss, structural similarity loss, Fourier loss, and histogram loss to ensure comprehensive and natural enhancement effects. The novelty of this algorithm lies not only in its multi-branch feature extraction design but also in its unique model structure, which synergistically improves image quality across different domains, effectively preventing over-enhancement, and ultimately achieving a balanced enhancement of brightness, details, and texture. Experimental results on multiple datasets, including SIDD, LOL, MIT-Adobe-FiveK, and LOL-v2-synthetic, demonstrate that the proposed method outperforms existing techniques in terms of image detail, texture, and brightness. Specifically, it achieves a PSNR of 27.52 dB on the LOL dataset, surpassing Wavelet Diffusion by 1.19 dB. Additionally, on the LOL-v2-synthetic dataset, it achieves a PSNR of 29.56 dB, exceeding Wavelet Diffusion by 3.06 dB. These results demonstrate a significant enhancement in the visual quality of low-light images.
期刊介绍:
Digital Signal Processing: A Review Journal is one of the oldest and most established journals in the field of signal processing yet it aims to be the most innovative. The Journal invites top quality research articles at the frontiers of research in all aspects of signal processing. Our objective is to provide a platform for the publication of ground-breaking research in signal processing with both academic and industrial appeal.
The journal has a special emphasis on statistical signal processing methodology such as Bayesian signal processing, and encourages articles on emerging applications of signal processing such as:
• big data• machine learning• internet of things• information security• systems biology and computational biology,• financial time series analysis,• autonomous vehicles,• quantum computing,• neuromorphic engineering,• human-computer interaction and intelligent user interfaces,• environmental signal processing,• geophysical signal processing including seismic signal processing,• chemioinformatics and bioinformatics,• audio, visual and performance arts,• disaster management and prevention,• renewable energy,