The proliferation of self-media and smart devices has led to uneven video quality on streaming platforms, so there is an urgent need for effective automated video quality assessment (VQA) methods. But most existing VQA methods fail to fully consider dynamic adaptability of the human visual perception system and its synergistic mechanism. In this study, we proposed a novel multi-path sensing framework for VQA to enhance the progressive sensing capability of the model. Specifically, the complete video has to be divided into three perceptual levels: patch clips, sampled frame stream, and inter-frame differences, a balance factor is used to give different levels of perceptual weights. Firstly, the purpose of defining a patch sampling method is to reduce the input data of the model while aligning temporal information, to extract subtle motion features in patch clips. After that, to further enhance the representation of local high-frequency details, the global variance-guided temporal dimension attention mechanism and spatial feature aggregation pool are used to accurately fit the sampling frame sequence. Finally, by embedding the feature map differences between consecutive frames and utilizing the long-term spatio-temporal dependence of Transformer to simulate the global dynamic evolution, the model achieves progressive interaction of cross scale spatio-temporal information. In addition, the improved temporal hysteresis pool enhances the ability to capture nonlinear dynamics in time series data and can more faithfully simulate subtle changes in the human visual perception system. Experimental results show that the proposed method outperforms existing no-reference VQA (NR-VQA) approaches across five in-the-wild datasets. In particular, it achieves outstanding performance on the CVD2014 dataset, which is the smallest in scale and contains the fewest scene variations, reaching a PLCC of 0.927 and an SRCC of 0.925. These results clearly demonstrate the effectiveness and advantages of our method in the VQA task.
扫码关注我们
求助内容:
应助结果提醒方式:
