Subjective and Objective Quality Assessment of Colonoscopy Videos

Guanghui Yue;Lixin Zhang;Jingfeng Du;Tianwei Zhou;Wei Zhou;Weisi Lin
{"title":"Subjective and Objective Quality Assessment of Colonoscopy Videos","authors":"Guanghui Yue;Lixin Zhang;Jingfeng Du;Tianwei Zhou;Wei Zhou;Weisi Lin","doi":"10.1109/TMI.2024.3461737","DOIUrl":null,"url":null,"abstract":"Captured colonoscopy videos usually suffer from multiple real-world distortions, such as motion blur, low brightness, abnormal exposure, and object occlusion, which impede visual interpretation. However, existing works mainly investigate the impacts of synthesized distortions, which differ from real-world distortions greatly. This research aims to carry out an in-depth study for colonoscopy Video Quality Assessment (VQA). In this study, we advance this topic by establishing both subjective and objective solutions. Firstly, we collect 1,000 colonoscopy videos with typical visual quality degradation conditions in practice and construct a multi-attribute VQA database. The quality of each video is annotated by subjective experiments from five distortion attributes (i.e., temporal-spatial visibility, brightness, specular reflection, stability, and utility), as well as an overall perspective. Secondly, we propose a Distortion Attribute Reasoning Network (DARNet) for automatic VQA. DARNet includes two streams to extract features related to spatial and temporal distortions, respectively. It adaptively aggregates the attribute-related features through a multi-attribute association module to predict the quality score of each distortion attribute. Motivated by the observation that the rating behaviors for all attributes are different, a behavior guided reasoning module is further used to fuse the attribute-aware features, resulting in the overall quality. Experimental results on the constructed database show that our DARNet correlates well with subjective ratings and is superior to nine state-of-the-art methods.","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"44 2","pages":"841-854"},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on medical imaging","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10681143/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Captured colonoscopy videos usually suffer from multiple real-world distortions, such as motion blur, low brightness, abnormal exposure, and object occlusion, which impede visual interpretation. However, existing works mainly investigate the impacts of synthesized distortions, which differ from real-world distortions greatly. This research aims to carry out an in-depth study for colonoscopy Video Quality Assessment (VQA). In this study, we advance this topic by establishing both subjective and objective solutions. Firstly, we collect 1,000 colonoscopy videos with typical visual quality degradation conditions in practice and construct a multi-attribute VQA database. The quality of each video is annotated by subjective experiments from five distortion attributes (i.e., temporal-spatial visibility, brightness, specular reflection, stability, and utility), as well as an overall perspective. Secondly, we propose a Distortion Attribute Reasoning Network (DARNet) for automatic VQA. DARNet includes two streams to extract features related to spatial and temporal distortions, respectively. It adaptively aggregates the attribute-related features through a multi-attribute association module to predict the quality score of each distortion attribute. Motivated by the observation that the rating behaviors for all attributes are different, a behavior guided reasoning module is further used to fuse the attribute-aware features, resulting in the overall quality. Experimental results on the constructed database show that our DARNet correlates well with subjective ratings and is superior to nine state-of-the-art methods.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
结肠镜检查视频的主观和客观质量评估
捕获的结肠镜检查视频通常会遭受多种现实世界的扭曲,例如运动模糊,低亮度,异常曝光和物体遮挡,这些都会妨碍视觉解释。然而,现有的研究主要是研究合成失真的影响,与真实失真有很大的不同。本研究旨在对结肠镜检查视频质量评估(VQA)进行深入研究。在本研究中,我们通过建立主观和客观的解决方案来推进这一主题。首先,我们收集了1000个在实践中具有典型视觉质量下降条件的结肠镜检查视频,构建了一个多属性的VQA数据库。每个视频的质量都是通过主观实验从五个失真属性(即时空可视性,亮度,镜面反射,稳定性和实用性)以及整体视角来注释的。其次,提出了一种用于自动VQA的失真属性推理网络(DARNet)。DARNet包括两个流,分别提取与空间和时间扭曲相关的特征。通过多属性关联模块自适应地聚合属性相关特征,预测每个失真属性的质量分数。由于观察到所有属性的评分行为不同,我们进一步使用行为引导推理模块来融合属性感知特征,从而提高整体质量。在构建的数据库上的实验结果表明,我们的DARNet与主观评分有很好的相关性,并且优于9种最先进的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
A 3D Cross-modal Keypoint Descriptor for MR-US Matching and Registration. FairFedMed: Benchmarking Group Fairness in Federated Medical Imaging With FairLoRA. MIRROR: Multi-Modal Pathological Self-Supervised Representation Learning via Modality Alignment and Retention Extractive Radiology Reporting With Memory-Based Cross-Modal Representations Wave-Aware Weakly Supervised Histopathological Tissue Segmentation With Cross-Scale Logits Distillation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1