Image Quality Assessment (IQA) is a fundamental task in computer vision, where existing methods often achieve superior performance by combining global and local representations. Inspired by the mechanism of human visual perception, where focus is placed on visually salient regions when assessing image quality, some studies have attempted to incorporate saliency information as a local feature to assist in quality prediction. However, these methods generally overlook the potential interaction between salient and background regions. To address this issue, we propose a novel IQA method, the Saliency Region Interaction Network (SRINet), which includes saliency-guided feature separation and encoding, region interaction enhancement, and multi-branch fusion. Specifically, image features are first partitioned into salient and background regions using a saliency mask, and each is embedded separately. A regional multi-head attention mechanism is then designed to model the interactive dependencies between these regions. Finally, a cross-attention mechanism, guided by the salient interaction features, fuses this local interactive information with global features, forming a comprehensive, quality-aware model. Experimental results on seven IQA databases demonstrate the competitiveness of SRINet on both synthetically and authentically distorted images.
扫码关注我们
求助内容:
应助结果提醒方式:
