Infrared and Visible Image Fusion (IVIF) is a technique used to integrate thermal information from infrared images with the fine details and textures of visible images to achieve comprehensive scene perception. It has broad applications in night vision, surveillance, and autonomous driving, where highlighting thermal targets while retaining scene details is essential. However, existing methods often struggle to effectively preserve modality-specific features and suffer from limited nonlinear modeling capacity, which hinders their ability to fully exploit the complementary information across modalities. In this research, we proposed KANFuse, a novel fusion network in which Kolmogorov–Arnold Networks (KAN) are incorporated to model complex nonlinear cross-modal interactions. To further enhance representation, Wavelet Convolution Blocks (WCBs) are employed for edge-aware and noise-suppressing feature extraction, while Dynamic Fusion Modules (DFMs) are integrated into skip connections to balance multi-source contributions. Additionally, a spectral guided fidelity loss (SFL) is designed for second-phase training to better retain realistic visual information. Extensive evaluations on TNO, , and LLVIP demonstrate that KANFuse consistently outperforms fourteen state-of-the-art methods in both qualitative and quantitative evaluations.
扫码关注我们
求助内容:
应助结果提醒方式:
