Nasim Jamshidi Avanaki, Abhijay Ghildiyal, Nabajeet Barman, Saman Zadtootaghaj
{"title":"MSLIQA: Enhancing Learning Representations for Image Quality Assessment through Multi-Scale Learning","authors":"Nasim Jamshidi Avanaki, Abhijay Ghildiyal, Nabajeet Barman, Saman Zadtootaghaj","doi":"arxiv-2408.16879","DOIUrl":null,"url":null,"abstract":"No-Reference Image Quality Assessment (NR-IQA) remains a challenging task due\nto the diversity of distortions and the lack of large annotated datasets. Many\nstudies have attempted to tackle these challenges by developing more accurate\nNR-IQA models, often employing complex and computationally expensive networks,\nor by bridging the domain gap between various distortions to enhance\nperformance on test datasets. In our work, we improve the performance of a\ngeneric lightweight NR-IQA model by introducing a novel augmentation strategy\nthat boosts its performance by almost 28\\%. This augmentation strategy enables\nthe network to better discriminate between different distortions in various\nparts of the image by zooming in and out. Additionally, the inclusion of\ntest-time augmentation further enhances performance, making our lightweight\nnetwork's results comparable to the current state-of-the-art models, simply\nthrough the use of augmentations.","PeriodicalId":501480,"journal":{"name":"arXiv - CS - Multimedia","volume":"9 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Multimedia","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.16879","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
No-Reference Image Quality Assessment (NR-IQA) remains a challenging task due
to the diversity of distortions and the lack of large annotated datasets. Many
studies have attempted to tackle these challenges by developing more accurate
NR-IQA models, often employing complex and computationally expensive networks,
or by bridging the domain gap between various distortions to enhance
performance on test datasets. In our work, we improve the performance of a
generic lightweight NR-IQA model by introducing a novel augmentation strategy
that boosts its performance by almost 28\%. This augmentation strategy enables
the network to better discriminate between different distortions in various
parts of the image by zooming in and out. Additionally, the inclusion of
test-time augmentation further enhances performance, making our lightweight
network's results comparable to the current state-of-the-art models, simply
through the use of augmentations.