Leveraging more information for blind stereo super-resolution via large-window cross-attention

IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Applied Soft Computing Pub Date : 2024-11-23 DOI:10.1016/j.asoc.2024.112492
Weifeng Cao, Xiaoyan Lei, Yong Jiang, Zongfei Bai, Xiaoliang Qian
{"title":"Leveraging more information for blind stereo super-resolution via large-window cross-attention","authors":"Weifeng Cao,&nbsp;Xiaoyan Lei,&nbsp;Yong Jiang,&nbsp;Zongfei Bai,&nbsp;Xiaoliang Qian","doi":"10.1016/j.asoc.2024.112492","DOIUrl":null,"url":null,"abstract":"<div><div>Stereo image super-resolution aims to reconstruct high-resolution images by effectively utilizing cross-view complementary information from stereo image pairs. A prevalent method in Stereo image super-resolution is the stereo cross-attention module, which allows the model to focus on and integrate relevant features from both the left and right views. Despite its advantages, our analysis using a diagnostic tool called local attribution map (LAM) reveals that current methods exhibit limitations in effectively leveraging this complementary information. To address this issue, we propose the Double Stereo Cross-Attention Module (DSCAM), which utilizes an Overlapping Stereo Cross-Attention (OSCA) mechanism that enhances the integration of cross-view complementary information by using overlapping windows, followed by an additional multiplication step to refine and emphasize the combined features. Additionally, we develop a stereo image degradation model that ensures the consistency of degradation between stereo pairs, accurately simulating the real-world degradation process of stereo images. Extensive experiments have demonstrated that our method achieves visually pleasing results, making it the first to address the problem of stereo image super-resolution in real-world scenarios. The source code is available at <span><span>https://github.com/nathan66666/LCASSR.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":"168 ","pages":"Article 112492"},"PeriodicalIF":7.2000,"publicationDate":"2024-11-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1568494624012663","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Stereo image super-resolution aims to reconstruct high-resolution images by effectively utilizing cross-view complementary information from stereo image pairs. A prevalent method in Stereo image super-resolution is the stereo cross-attention module, which allows the model to focus on and integrate relevant features from both the left and right views. Despite its advantages, our analysis using a diagnostic tool called local attribution map (LAM) reveals that current methods exhibit limitations in effectively leveraging this complementary information. To address this issue, we propose the Double Stereo Cross-Attention Module (DSCAM), which utilizes an Overlapping Stereo Cross-Attention (OSCA) mechanism that enhances the integration of cross-view complementary information by using overlapping windows, followed by an additional multiplication step to refine and emphasize the combined features. Additionally, we develop a stereo image degradation model that ensures the consistency of degradation between stereo pairs, accurately simulating the real-world degradation process of stereo images. Extensive experiments have demonstrated that our method achieves visually pleasing results, making it the first to address the problem of stereo image super-resolution in real-world scenarios. The source code is available at https://github.com/nathan66666/LCASSR.git.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
通过大窗口交叉关注,利用更多信息实现盲立体超分辨率
立体图像超分辨率旨在通过有效利用立体图像对的跨视角互补信息来重建高分辨率图像。立体图像超分辨率的一种常用方法是立体交叉关注模块,它允许模型关注并整合来自左右视图的相关特征。尽管这种方法有很多优点,但我们使用一种名为局部归因图(LAM)的诊断工具进行的分析表明,目前的方法在有效利用这些互补信息方面存在局限性。为了解决这个问题,我们提出了双立体交叉注意模块(DSCAM),该模块利用重叠立体交叉注意(OSCA)机制,通过使用重叠窗口来加强跨视角互补信息的整合,然后通过额外的乘法步骤来完善和强调组合特征。此外,我们还开发了一种立体图像降解模型,可确保立体图像对之间降解的一致性,准确模拟现实世界中立体图像的降解过程。广泛的实验证明,我们的方法取得了令人满意的视觉效果,首次解决了现实世界中的立体图像超分辨率问题。源代码见 https://github.com/nathan66666/LCASSR.git。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Applied Soft Computing
Applied Soft Computing 工程技术-计算机:跨学科应用
CiteScore
15.80
自引率
6.90%
发文量
874
审稿时长
10.9 months
期刊介绍: Applied Soft Computing is an international journal promoting an integrated view of soft computing to solve real life problems.The focus is to publish the highest quality research in application and convergence of the areas of Fuzzy Logic, Neural Networks, Evolutionary Computing, Rough Sets and other similar techniques to address real world complexities. Applied Soft Computing is a rolling publication: articles are published as soon as the editor-in-chief has accepted them. Therefore, the web site will continuously be updated with new articles and the publication time will be short.
期刊最新文献
Editorial Board Editorial Board Adolescent mental health state assessment framework by combining YOLO with random forest Leveraging more information for blind stereo super-resolution via large-window cross-attention Optimized hybrid XGBoost-CatBoost model for enhanced prediction of concrete strength and reliability analysis using Monte Carlo simulations
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1