FGR-Net:基于深度重构学习的可解释眼底图像降级性分类

Saif Khalid, Hatem A. Rashwan, Saddam Abdulwahab, Mohamed Abdel-Nasser, Facundo Manuel Quiroga, Domenec Puig
{"title":"FGR-Net:基于深度重构学习的可解释眼底图像降级性分类","authors":"Saif Khalid, Hatem A. Rashwan, Saddam Abdulwahab, Mohamed Abdel-Nasser, Facundo Manuel Quiroga, Domenec Puig","doi":"arxiv-2409.10246","DOIUrl":null,"url":null,"abstract":"The performance of diagnostic Computer-Aided Design (CAD) systems for retinal\ndiseases depends on the quality of the retinal images being screened. Thus,\nmany studies have been developed to evaluate and assess the quality of such\nretinal images. However, most of them did not investigate the relationship\nbetween the accuracy of the developed models and the quality of the\nvisualization of interpretability methods for distinguishing between gradable\nand non-gradable retinal images. Consequently, this paper presents a novel\nframework called FGR-Net to automatically assess and interpret underlying\nfundus image quality by merging an autoencoder network with a classifier\nnetwork. The FGR-Net model also provides an interpretable quality assessment\nthrough visualizations. In particular, FGR-Net uses a deep autoencoder to\nreconstruct the input image in order to extract the visual characteristics of\nthe input fundus images based on self-supervised learning. The extracted\nfeatures by the autoencoder are then fed into a deep classifier network to\ndistinguish between gradable and ungradable fundus images. FGR-Net is evaluated\nwith different interpretability methods, which indicates that the autoencoder\nis a key factor in forcing the classifier to focus on the relevant structures\nof the fundus images, such as the fovea, optic disk, and prominent blood\nvessels. Additionally, the interpretability methods can provide visual feedback\nfor ophthalmologists to understand how our model evaluates the quality of\nfundus images. The experimental results showed the superiority of FGR-Net over\nthe state-of-the-art quality assessment methods, with an accuracy of 89% and an\nF1-score of 87%.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"FGR-Net:Interpretable fundus imagegradeability classification based on deepreconstruction learning\",\"authors\":\"Saif Khalid, Hatem A. Rashwan, Saddam Abdulwahab, Mohamed Abdel-Nasser, Facundo Manuel Quiroga, Domenec Puig\",\"doi\":\"arxiv-2409.10246\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The performance of diagnostic Computer-Aided Design (CAD) systems for retinal\\ndiseases depends on the quality of the retinal images being screened. Thus,\\nmany studies have been developed to evaluate and assess the quality of such\\nretinal images. However, most of them did not investigate the relationship\\nbetween the accuracy of the developed models and the quality of the\\nvisualization of interpretability methods for distinguishing between gradable\\nand non-gradable retinal images. Consequently, this paper presents a novel\\nframework called FGR-Net to automatically assess and interpret underlying\\nfundus image quality by merging an autoencoder network with a classifier\\nnetwork. The FGR-Net model also provides an interpretable quality assessment\\nthrough visualizations. In particular, FGR-Net uses a deep autoencoder to\\nreconstruct the input image in order to extract the visual characteristics of\\nthe input fundus images based on self-supervised learning. The extracted\\nfeatures by the autoencoder are then fed into a deep classifier network to\\ndistinguish between gradable and ungradable fundus images. FGR-Net is evaluated\\nwith different interpretability methods, which indicates that the autoencoder\\nis a key factor in forcing the classifier to focus on the relevant structures\\nof the fundus images, such as the fovea, optic disk, and prominent blood\\nvessels. Additionally, the interpretability methods can provide visual feedback\\nfor ophthalmologists to understand how our model evaluates the quality of\\nfundus images. The experimental results showed the superiority of FGR-Net over\\nthe state-of-the-art quality assessment methods, with an accuracy of 89% and an\\nF1-score of 87%.\",\"PeriodicalId\":501289,\"journal\":{\"name\":\"arXiv - EE - Image and Video Processing\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - EE - Image and Video Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.10246\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - EE - Image and Video Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10246","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

视网膜疾病诊断计算机辅助设计(CAD)系统的性能取决于所筛选视网膜图像的质量。因此,许多研究都对此类视网膜图像的质量进行了评估和评价。然而,其中大多数研究并没有调查所开发模型的准确性与区分可分级和不可分级视网膜图像的可视化可解释性方法的质量之间的关系。因此,本文提出了一种名为 FGR-Net 的新型框架,通过合并自动编码器网络和分类网络来自动评估和解释底层视网膜图像质量。FGR-Net 模型还通过可视化提供了可解释的质量评估。特别是,FGR-Net 使用深度自动编码器重新构建输入图像,以便在自我监督学习的基础上提取输入眼底图像的视觉特征。然后将自动编码器提取的特征输入深度分类器网络,以区分可渐变和不可渐变的眼底图像。使用不同的可解释性方法对 FGR-Net 进行了评估,结果表明,自动编码器是迫使分类器关注眼底图像相关结构(如眼窝、视盘和突出的血管)的关键因素。此外,可解释性方法还能为眼科医生提供视觉反馈,让他们了解我们的模型是如何评估眼底图像质量的。实验结果表明,FGR-Net 的准确率为 89%,F1 分数为 87%,优于最先进的质量评估方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
FGR-Net:Interpretable fundus imagegradeability classification based on deepreconstruction learning
The performance of diagnostic Computer-Aided Design (CAD) systems for retinal diseases depends on the quality of the retinal images being screened. Thus, many studies have been developed to evaluate and assess the quality of such retinal images. However, most of them did not investigate the relationship between the accuracy of the developed models and the quality of the visualization of interpretability methods for distinguishing between gradable and non-gradable retinal images. Consequently, this paper presents a novel framework called FGR-Net to automatically assess and interpret underlying fundus image quality by merging an autoencoder network with a classifier network. The FGR-Net model also provides an interpretable quality assessment through visualizations. In particular, FGR-Net uses a deep autoencoder to reconstruct the input image in order to extract the visual characteristics of the input fundus images based on self-supervised learning. The extracted features by the autoencoder are then fed into a deep classifier network to distinguish between gradable and ungradable fundus images. FGR-Net is evaluated with different interpretability methods, which indicates that the autoencoder is a key factor in forcing the classifier to focus on the relevant structures of the fundus images, such as the fovea, optic disk, and prominent blood vessels. Additionally, the interpretability methods can provide visual feedback for ophthalmologists to understand how our model evaluates the quality of fundus images. The experimental results showed the superiority of FGR-Net over the state-of-the-art quality assessment methods, with an accuracy of 89% and an F1-score of 87%.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
multiPI-TransBTS: A Multi-Path Learning Framework for Brain Tumor Image Segmentation Based on Multi-Physical Information Autopet III challenge: Incorporating anatomical knowledge into nnUNet for lesion segmentation in PET/CT Denoising diffusion models for high-resolution microscopy image restoration Tumor aware recurrent inter-patient deformable image registration of computed tomography scans with lung cancer Cross-Organ and Cross-Scanner Adenocarcinoma Segmentation using Rein to Fine-tune Vision Foundation Models
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1