{"title":"用于图像超分辨率的分频启始变压器","authors":"Wei Xu","doi":"10.1145/3609703.3609708","DOIUrl":null,"url":null,"abstract":"Transformer models have shown remarkable effectiveness in capturing long-range dependencies and extracting features for single image super-resolution. However, their deployment on edge devices is hindered by their high computational complexity. To address this challenge, we propose Inception Swin Transformer (IST), a novel model that leverages frequency domain separation to reduce redundant computations.In IST, we exploit the strengths of both CNN-based networks and Transformer variants to handle high-frequency and low-frequency features, respectively. By dynamically utilizing frequency factors to separate feature maps, IST ensures that different components are processed appropriately. Additionally, IST maintains a balanced trade-off between model speed and performance by gradually reducing the proportion of high-frequency components.Our experiments demonstrate that IST effectively reduces the FLOPs while preserving high performance. The combination of Transformers’ accuracy and CNN variants’ efficiency enables IST to significantly reduce computational strain without compromising quality. Comparative analysis reveals that IST outperforms other models, achieving superior results with less FLOPs.","PeriodicalId":101485,"journal":{"name":"Proceedings of the 2023 5th International Conference on Pattern Recognition and Intelligent Systems","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Frequency-Split Inception Transformer for Image Super-Resolution\",\"authors\":\"Wei Xu\",\"doi\":\"10.1145/3609703.3609708\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Transformer models have shown remarkable effectiveness in capturing long-range dependencies and extracting features for single image super-resolution. However, their deployment on edge devices is hindered by their high computational complexity. To address this challenge, we propose Inception Swin Transformer (IST), a novel model that leverages frequency domain separation to reduce redundant computations.In IST, we exploit the strengths of both CNN-based networks and Transformer variants to handle high-frequency and low-frequency features, respectively. By dynamically utilizing frequency factors to separate feature maps, IST ensures that different components are processed appropriately. Additionally, IST maintains a balanced trade-off between model speed and performance by gradually reducing the proportion of high-frequency components.Our experiments demonstrate that IST effectively reduces the FLOPs while preserving high performance. The combination of Transformers’ accuracy and CNN variants’ efficiency enables IST to significantly reduce computational strain without compromising quality. Comparative analysis reveals that IST outperforms other models, achieving superior results with less FLOPs.\",\"PeriodicalId\":101485,\"journal\":{\"name\":\"Proceedings of the 2023 5th International Conference on Pattern Recognition and Intelligent Systems\",\"volume\":\"25 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-07-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2023 5th International Conference on Pattern Recognition and Intelligent Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3609703.3609708\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2023 5th International Conference on Pattern Recognition and Intelligent Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3609703.3609708","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Frequency-Split Inception Transformer for Image Super-Resolution
Transformer models have shown remarkable effectiveness in capturing long-range dependencies and extracting features for single image super-resolution. However, their deployment on edge devices is hindered by their high computational complexity. To address this challenge, we propose Inception Swin Transformer (IST), a novel model that leverages frequency domain separation to reduce redundant computations.In IST, we exploit the strengths of both CNN-based networks and Transformer variants to handle high-frequency and low-frequency features, respectively. By dynamically utilizing frequency factors to separate feature maps, IST ensures that different components are processed appropriately. Additionally, IST maintains a balanced trade-off between model speed and performance by gradually reducing the proportion of high-frequency components.Our experiments demonstrate that IST effectively reduces the FLOPs while preserving high performance. The combination of Transformers’ accuracy and CNN variants’ efficiency enables IST to significantly reduce computational strain without compromising quality. Comparative analysis reveals that IST outperforms other models, achieving superior results with less FLOPs.