Fusion of infrared and visible images via multi-layer convolutional sparse representation

IF 5.2 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Journal of King Saud University-Computer and Information Sciences Pub Date : 2024-07-01 DOI:10.1016/j.jksuci.2024.102090
Zhouyu Zhang , Chenyuan He , Hai Wang , Yingfeng Cai , Long Chen , Zhihua Gan , Fenghua Huang , Yiqun Zhang
{"title":"Fusion of infrared and visible images via multi-layer convolutional sparse representation","authors":"Zhouyu Zhang ,&nbsp;Chenyuan He ,&nbsp;Hai Wang ,&nbsp;Yingfeng Cai ,&nbsp;Long Chen ,&nbsp;Zhihua Gan ,&nbsp;Fenghua Huang ,&nbsp;Yiqun Zhang","doi":"10.1016/j.jksuci.2024.102090","DOIUrl":null,"url":null,"abstract":"<div><p>Infrared and visible image fusion is an effective solution for image quality enhancement. However, conventional fusion models require the decomposition of source images into image blocks, which disrupts the original structure of the images, leading to the loss of detail in the fused images and making the fusion results highly sensitive to matching errors. This paper employs Convolutional Sparse Representation (CSR) to perform global feature transformation on the source images, overcoming the drawbacks of traditional fusion models that rely on image decomposition. Inspired by neural networks, a multi-layer CSR model is proposed, which involves five layers in a forward-feeding manner: two CSR layers acquiring sparse coefficient maps, one fusion layer combining sparse maps, and two reconstruction layers for image recovery. The dataset used in this paper comprises infrared and visible images selected from public dataset, as well as registered images collected by an actual Unmanned Aerial Vehicle (UAV). The source images contain ground targets, marine targets, and natural landscapes. To validate the effectiveness of the proposed image fusion model in this paper, comparative analysis is conducted with state-of-the-art (SOTA) algorithms. Experimental results demonstrate that the proposed fusion model outperforms other state-of-the-art methods by at least 10% in SF, EN, MI and <span><math><msup><mrow><mi>Q</mi></mrow><mrow><mi>A</mi><mi>B</mi><mo>/</mo><mi>F</mi></mrow></msup></math></span> fusion metrics in most image fusion cases, thereby affirming its favorable performance.</p></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":null,"pages":null},"PeriodicalIF":5.2000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1319157824001794/pdfft?md5=519b5bf350ebfdc5c76e12245ab0600b&pid=1-s2.0-S1319157824001794-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of King Saud University-Computer and Information Sciences","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1319157824001794","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Infrared and visible image fusion is an effective solution for image quality enhancement. However, conventional fusion models require the decomposition of source images into image blocks, which disrupts the original structure of the images, leading to the loss of detail in the fused images and making the fusion results highly sensitive to matching errors. This paper employs Convolutional Sparse Representation (CSR) to perform global feature transformation on the source images, overcoming the drawbacks of traditional fusion models that rely on image decomposition. Inspired by neural networks, a multi-layer CSR model is proposed, which involves five layers in a forward-feeding manner: two CSR layers acquiring sparse coefficient maps, one fusion layer combining sparse maps, and two reconstruction layers for image recovery. The dataset used in this paper comprises infrared and visible images selected from public dataset, as well as registered images collected by an actual Unmanned Aerial Vehicle (UAV). The source images contain ground targets, marine targets, and natural landscapes. To validate the effectiveness of the proposed image fusion model in this paper, comparative analysis is conducted with state-of-the-art (SOTA) algorithms. Experimental results demonstrate that the proposed fusion model outperforms other state-of-the-art methods by at least 10% in SF, EN, MI and QAB/F fusion metrics in most image fusion cases, thereby affirming its favorable performance.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
通过多层卷积稀疏表示法融合红外和可见光图像
红外和可见光图像融合是提高图像质量的有效解决方案。然而,传统的融合模型需要将源图像分解成图像块,这破坏了图像的原始结构,导致融合后的图像细节丢失,使融合结果对匹配误差高度敏感。本文采用卷积稀疏表示法(CSR)对源图像进行全局特征变换,克服了传统融合模型依赖图像分解的缺点。受神经网络的启发,我们提出了一种多层 CSR 模型,它以前馈方式包含五个层:两个获取稀疏系数图的 CSR 层,一个结合稀疏图的融合层,以及两个用于图像复原的重建层。本文使用的数据集包括从公共数据集中选取的红外和可见光图像,以及实际无人飞行器(UAV)采集的注册图像。源图像包括地面目标、海洋目标和自然景观。为了验证本文提出的图像融合模型的有效性,我们与最先进的(SOTA)算法进行了对比分析。实验结果表明,在大多数图像融合情况下,所提出的融合模型在 SF、EN、MI 和 QAB/F 融合指标上至少比其他先进方法高出 10%,从而肯定了其良好的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
10.50
自引率
8.70%
发文量
656
审稿时长
29 days
期刊介绍: In 2022 the Journal of King Saud University - Computer and Information Sciences will become an author paid open access journal. Authors who submit their manuscript after October 31st 2021 will be asked to pay an Article Processing Charge (APC) after acceptance of their paper to make their work immediately, permanently, and freely accessible to all. The Journal of King Saud University Computer and Information Sciences is a refereed, international journal that covers all aspects of both foundations of computer and its practical applications.
期刊最新文献
Heterogeneous emotional contagion of the cyber–physical society A novel edge intelligence-based solution for safer footpath navigation of visually impaired using computer vision Improving embedding-based link prediction performance using clustering A sharding blockchain protocol for enhanced scalability and performance optimization through account transaction reconfiguration RAPID: Robust multi-pAtch masker using channel-wise Pooled varIance with two-stage patch Detection
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1