Jinsheng Fang, Hanjiang Lin, Jianglong Zhao, Kun Zeng
{"title":"用于轻量级图像超分辨率的高效多尺度大型非对称核网络","authors":"Jinsheng Fang, Hanjiang Lin, Jianglong Zhao, Kun Zeng","doi":"10.1002/cpe.8240","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>Recently, lightweight convolutional neural networks (CNNs) on single image super-resolution (SISR) tasks have received impressive improvement with delicate structures. However, numerous lightweight methods may reduce the representation capacity of the network due to decreasing the model size and computational complexities, leading to unsatisfactory performance. In this paper, we propose an efficient multi-scale large asymmetric-kernel network (MLAN) for lightweight SISR. Specifically, MLAN is built with a succession of feature cross extraction blocks (FCEBs), which better models local and long-range interactive information of features for SR. Each of the FCEB contains a multi-scale asymmetric large-kernel attention block (MACAB) by using multiple convolutional kernels to extract features in different receptive fields and a gated mechanism to preserve the useful information for SR. Extensive experimental results on five public benchmark datasets demonstrate the superiority of MLAN over the other advanced lightweight SISR competitors. The average PSNR values are about 0.12, 0.17 and 0.11 dB greater than the second-best competitors under scaling factors of ×2, ×3 and ×4, respectively. The proposed efficient blocks enable our MLAN to make a better balance between model size and performance and achieve comparable performance with Transformer-based methods at a similar level of parameters.</p>\n </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"36 23","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An efficient multi-scale large asymmetric-kernel network for lightweight image super-resolution\",\"authors\":\"Jinsheng Fang, Hanjiang Lin, Jianglong Zhao, Kun Zeng\",\"doi\":\"10.1002/cpe.8240\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n <p>Recently, lightweight convolutional neural networks (CNNs) on single image super-resolution (SISR) tasks have received impressive improvement with delicate structures. However, numerous lightweight methods may reduce the representation capacity of the network due to decreasing the model size and computational complexities, leading to unsatisfactory performance. In this paper, we propose an efficient multi-scale large asymmetric-kernel network (MLAN) for lightweight SISR. Specifically, MLAN is built with a succession of feature cross extraction blocks (FCEBs), which better models local and long-range interactive information of features for SR. Each of the FCEB contains a multi-scale asymmetric large-kernel attention block (MACAB) by using multiple convolutional kernels to extract features in different receptive fields and a gated mechanism to preserve the useful information for SR. Extensive experimental results on five public benchmark datasets demonstrate the superiority of MLAN over the other advanced lightweight SISR competitors. The average PSNR values are about 0.12, 0.17 and 0.11 dB greater than the second-best competitors under scaling factors of ×2, ×3 and ×4, respectively. The proposed efficient blocks enable our MLAN to make a better balance between model size and performance and achieve comparable performance with Transformer-based methods at a similar level of parameters.</p>\\n </div>\",\"PeriodicalId\":55214,\"journal\":{\"name\":\"Concurrency and Computation-Practice & Experience\",\"volume\":\"36 23\",\"pages\":\"\"},\"PeriodicalIF\":1.5000,\"publicationDate\":\"2024-07-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Concurrency and Computation-Practice & Experience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/cpe.8240\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, SOFTWARE ENGINEERING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Concurrency and Computation-Practice & Experience","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/cpe.8240","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
An efficient multi-scale large asymmetric-kernel network for lightweight image super-resolution
Recently, lightweight convolutional neural networks (CNNs) on single image super-resolution (SISR) tasks have received impressive improvement with delicate structures. However, numerous lightweight methods may reduce the representation capacity of the network due to decreasing the model size and computational complexities, leading to unsatisfactory performance. In this paper, we propose an efficient multi-scale large asymmetric-kernel network (MLAN) for lightweight SISR. Specifically, MLAN is built with a succession of feature cross extraction blocks (FCEBs), which better models local and long-range interactive information of features for SR. Each of the FCEB contains a multi-scale asymmetric large-kernel attention block (MACAB) by using multiple convolutional kernels to extract features in different receptive fields and a gated mechanism to preserve the useful information for SR. Extensive experimental results on five public benchmark datasets demonstrate the superiority of MLAN over the other advanced lightweight SISR competitors. The average PSNR values are about 0.12, 0.17 and 0.11 dB greater than the second-best competitors under scaling factors of ×2, ×3 and ×4, respectively. The proposed efficient blocks enable our MLAN to make a better balance between model size and performance and achieve comparable performance with Transformer-based methods at a similar level of parameters.
期刊介绍:
Concurrency and Computation: Practice and Experience (CCPE) publishes high-quality, original research papers, and authoritative research review papers, in the overlapping fields of:
Parallel and distributed computing;
High-performance computing;
Computational and data science;
Artificial intelligence and machine learning;
Big data applications, algorithms, and systems;
Network science;
Ontologies and semantics;
Security and privacy;
Cloud/edge/fog computing;
Green computing; and
Quantum computing.