{"title":"MAFBLiF:基于多尺度注意力特征融合的盲光场图像质量评估","authors":"Rui Zhou;Gangyi Jiang;Yueli Cui;Yeyao Chen;Haiyong Xu;Ting Luo;Mei Yu","doi":"10.1109/TBC.2024.3434699","DOIUrl":null,"url":null,"abstract":"Light field imaging captures both the intensity and directional information of light rays, providing users with more immersive visual experience. However, during the processes of imaging, processing, coding and reconstruction, light field images (LFIs) may encounter various distortions that degrade their visual quality. Compared to two-dimensional image quality assessment, light field image quality assessment (LFIQA) needs to consider not only the image quality in the spatial domain but also the quality degradation in the angular domain. To effectively model the factors related to visual perception and LFI quality, this paper proposes a multi-scale attention feature fusion based blind LFIQA metric, named MAFBLiF. The proposed metric consists of the following parts: MLI-Patch generation, spatial-angular feature separation module, spatial-angular feature extraction backbone network, pyramid feature alignment module and patch attention module. These modules are specifically designed to extract spatial and angular information of LFIs, and capture multi-level information and regions of interest. Furthermore, a pooling scheme guided by the LFI’s gradient information and saliency is proposed, which integrates the quality of all MLI-patches into the overall quality of the input LFI. Finally, to demonstrate the effectiveness of the proposed metric, extensive experiments are conducted on three representative LFI quality evaluation datasets. The experimental results show that the proposed metric outperforms other state-of-the-art image quality assessment metrics. The code will be publicly available at \n<uri>https://github.com/oldblackfish/MAFBLiF</uri>\n.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 4","pages":"1266-1278"},"PeriodicalIF":3.2000,"publicationDate":"2024-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"MAFBLiF: Multi-Scale Attention Feature Fusion-Based Blind Light Field Image Quality Assessment\",\"authors\":\"Rui Zhou;Gangyi Jiang;Yueli Cui;Yeyao Chen;Haiyong Xu;Ting Luo;Mei Yu\",\"doi\":\"10.1109/TBC.2024.3434699\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Light field imaging captures both the intensity and directional information of light rays, providing users with more immersive visual experience. However, during the processes of imaging, processing, coding and reconstruction, light field images (LFIs) may encounter various distortions that degrade their visual quality. Compared to two-dimensional image quality assessment, light field image quality assessment (LFIQA) needs to consider not only the image quality in the spatial domain but also the quality degradation in the angular domain. To effectively model the factors related to visual perception and LFI quality, this paper proposes a multi-scale attention feature fusion based blind LFIQA metric, named MAFBLiF. The proposed metric consists of the following parts: MLI-Patch generation, spatial-angular feature separation module, spatial-angular feature extraction backbone network, pyramid feature alignment module and patch attention module. These modules are specifically designed to extract spatial and angular information of LFIs, and capture multi-level information and regions of interest. Furthermore, a pooling scheme guided by the LFI’s gradient information and saliency is proposed, which integrates the quality of all MLI-patches into the overall quality of the input LFI. Finally, to demonstrate the effectiveness of the proposed metric, extensive experiments are conducted on three representative LFI quality evaluation datasets. The experimental results show that the proposed metric outperforms other state-of-the-art image quality assessment metrics. The code will be publicly available at \\n<uri>https://github.com/oldblackfish/MAFBLiF</uri>\\n.\",\"PeriodicalId\":13159,\"journal\":{\"name\":\"IEEE Transactions on Broadcasting\",\"volume\":\"70 4\",\"pages\":\"1266-1278\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2024-08-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Broadcasting\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10623345/\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Broadcasting","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10623345/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Light field imaging captures both the intensity and directional information of light rays, providing users with more immersive visual experience. However, during the processes of imaging, processing, coding and reconstruction, light field images (LFIs) may encounter various distortions that degrade their visual quality. Compared to two-dimensional image quality assessment, light field image quality assessment (LFIQA) needs to consider not only the image quality in the spatial domain but also the quality degradation in the angular domain. To effectively model the factors related to visual perception and LFI quality, this paper proposes a multi-scale attention feature fusion based blind LFIQA metric, named MAFBLiF. The proposed metric consists of the following parts: MLI-Patch generation, spatial-angular feature separation module, spatial-angular feature extraction backbone network, pyramid feature alignment module and patch attention module. These modules are specifically designed to extract spatial and angular information of LFIs, and capture multi-level information and regions of interest. Furthermore, a pooling scheme guided by the LFI’s gradient information and saliency is proposed, which integrates the quality of all MLI-patches into the overall quality of the input LFI. Finally, to demonstrate the effectiveness of the proposed metric, extensive experiments are conducted on three representative LFI quality evaluation datasets. The experimental results show that the proposed metric outperforms other state-of-the-art image quality assessment metrics. The code will be publicly available at
https://github.com/oldblackfish/MAFBLiF
.
期刊介绍:
The Society’s Field of Interest is “Devices, equipment, techniques and systems related to broadcast technology, including the production, distribution, transmission, and propagation aspects.” In addition to this formal FOI statement, which is used to provide guidance to the Publications Committee in the selection of content, the AdCom has further resolved that “broadcast systems includes all aspects of transmission, propagation, and reception.”