The objective of combining infrared with visible images lies in merging essential visual data from both sources to produce an enhanced output. Existing fusion methods predominantly operate within the spatial domain, while ignoring valuable data that could be extracted from the frequency domain. Therefore, the fusion performance remains suboptimal. To overcome this drawback, we introduce the Spatial-Frequency Edge-Aware Network(SFEANet) model, which employs a parallel dual-branch structure that simultaneously processes spatial and frequency domain information. The spatial fusion branch utilizes the Edge Feature Extraction(EFE) block and the Self Attention(SA) block to capture and integrate key features across both image types. The frequency-domain fusion branch first applies the Fast Fourier Transform(FFT) for domain conversion, which transforms the input into spectral representations. Subsequently, it performs interactive operations on their amplitude and phase components to enable cross-modal feature integration. The fused features are ultimately reconstructed in the spatial domain through the Inverse Fast Fourier Transform (IFFT). Comprehensive experiments conducted on three public benchmarks demonstrate the superior performance of SFEANet across multiple quantitative measures and perceptual quality assessments. The implementation can be accessed via https://github.com/lishuohui123/SFEANet.
扫码关注我们
求助内容:
应助结果提醒方式:
