ViT-AMD:从眼底图像诊断年龄相关性黄斑变性的新型深度学习模型

IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE International Journal of Intelligent Systems Pub Date : 2024-11-15 DOI:10.1155/2024/3026500
Ngoc Thien Le, Thanh Le Truong, Sunchai Deelertpaiboon, Wattanasak Srisiri, Pear Ferreira Pongsachareonnont, Disorn Suwajanakorn, Apivat Mavichak, Rath Itthipanichpong, Widhyakorn Asdornwised, Watit Benjapolakul, Surachai Chaitusaney, Pasu Kaewplung
{"title":"ViT-AMD:从眼底图像诊断年龄相关性黄斑变性的新型深度学习模型","authors":"Ngoc Thien Le,&nbsp;Thanh Le Truong,&nbsp;Sunchai Deelertpaiboon,&nbsp;Wattanasak Srisiri,&nbsp;Pear Ferreira Pongsachareonnont,&nbsp;Disorn Suwajanakorn,&nbsp;Apivat Mavichak,&nbsp;Rath Itthipanichpong,&nbsp;Widhyakorn Asdornwised,&nbsp;Watit Benjapolakul,&nbsp;Surachai Chaitusaney,&nbsp;Pasu Kaewplung","doi":"10.1155/2024/3026500","DOIUrl":null,"url":null,"abstract":"<div>\n <p>Age-related macular degeneration (AMD) diagnosis using fundus images is one of the critical missions of the eye-care screening program in many countries. Various proposed deep learning models have been studied for this research interest, which aim to achieve the mission and outperform human-based approaches. However, research efforts are still required for the improvement of model classification accuracy, sensitivity, and specificity values. In this study, we proposed the model named as ViT-AMD, which is based on the latest Vision Transformer (ViT) structure, to diagnosis a fundus image as normal, dry AMD, or wet AMD types. Unlike convolution neural network models, ViT consists of the attention map layers, which show more effective performance for image classification task. Our training process is based on the 5-fold cross-validation and transfer learning techniques using Chula-AMD dataset at the Department of Ophthalmology, the King Chulalongkorn Memorial Hospital, Bangkok. Furthermore, we also test the performance of trained model using an independent image datasets. The results showed that for the 3-classes AMD classification (normal vs. dry AMD vs. wet AMD) on the Chula-AMD dataset, the averaged accuracy, precision, sensitivity, and specificity of our trained model are about 93.40%, 92.15%, 91.27%, and 96.57%, respectively. For result testing on independent datasets, the averaged accuracy, precision, sensitivity, and specificity of trained model are about 74, 20%, 75.35%, 74.13%, and 87.07%, respectively. Compared with the results from the baseline CNN-based model (DenseNet201), the trained ViT-AMD model has outperformed significantly. In conclusion, the ViT-AMD model have proved their usefulness to assist the ophthalmologist to diagnosis the AMD disease.</p>\n </div>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":"2024 1","pages":""},"PeriodicalIF":5.0000,"publicationDate":"2024-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1155/2024/3026500","citationCount":"0","resultStr":"{\"title\":\"ViT-AMD: A New Deep Learning Model for Age-Related Macular Degeneration Diagnosis From Fundus Images\",\"authors\":\"Ngoc Thien Le,&nbsp;Thanh Le Truong,&nbsp;Sunchai Deelertpaiboon,&nbsp;Wattanasak Srisiri,&nbsp;Pear Ferreira Pongsachareonnont,&nbsp;Disorn Suwajanakorn,&nbsp;Apivat Mavichak,&nbsp;Rath Itthipanichpong,&nbsp;Widhyakorn Asdornwised,&nbsp;Watit Benjapolakul,&nbsp;Surachai Chaitusaney,&nbsp;Pasu Kaewplung\",\"doi\":\"10.1155/2024/3026500\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n <p>Age-related macular degeneration (AMD) diagnosis using fundus images is one of the critical missions of the eye-care screening program in many countries. Various proposed deep learning models have been studied for this research interest, which aim to achieve the mission and outperform human-based approaches. However, research efforts are still required for the improvement of model classification accuracy, sensitivity, and specificity values. In this study, we proposed the model named as ViT-AMD, which is based on the latest Vision Transformer (ViT) structure, to diagnosis a fundus image as normal, dry AMD, or wet AMD types. Unlike convolution neural network models, ViT consists of the attention map layers, which show more effective performance for image classification task. Our training process is based on the 5-fold cross-validation and transfer learning techniques using Chula-AMD dataset at the Department of Ophthalmology, the King Chulalongkorn Memorial Hospital, Bangkok. Furthermore, we also test the performance of trained model using an independent image datasets. The results showed that for the 3-classes AMD classification (normal vs. dry AMD vs. wet AMD) on the Chula-AMD dataset, the averaged accuracy, precision, sensitivity, and specificity of our trained model are about 93.40%, 92.15%, 91.27%, and 96.57%, respectively. For result testing on independent datasets, the averaged accuracy, precision, sensitivity, and specificity of trained model are about 74, 20%, 75.35%, 74.13%, and 87.07%, respectively. Compared with the results from the baseline CNN-based model (DenseNet201), the trained ViT-AMD model has outperformed significantly. In conclusion, the ViT-AMD model have proved their usefulness to assist the ophthalmologist to diagnosis the AMD disease.</p>\\n </div>\",\"PeriodicalId\":14089,\"journal\":{\"name\":\"International Journal of Intelligent Systems\",\"volume\":\"2024 1\",\"pages\":\"\"},\"PeriodicalIF\":5.0000,\"publicationDate\":\"2024-11-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1155/2024/3026500\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Intelligent Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1155/2024/3026500\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1155/2024/3026500","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

利用眼底图像诊断老年性黄斑变性(AMD)是许多国家眼科筛查项目的重要任务之一。针对这一研究兴趣,人们研究了各种拟议的深度学习模型,旨在实现这一任务并超越基于人类的方法。然而,要提高模型的分类准确性、灵敏度和特异性值,仍需努力研究。在本研究中,我们提出了基于最新视觉转换器(ViT)结构的 ViT-AMD 模型,以诊断眼底图像为正常、干性 AMD 或湿性 AMD 类型。与卷积神经网络模型不同,ViT 由注意力图层组成,在图像分类任务中表现出更有效的性能。我们利用曼谷朱拉隆功国王纪念医院眼科部的 Chula-AMD 数据集,采用 5 倍交叉验证和迁移学习技术进行训练。此外,我们还使用独立的图像数据集测试了训练模型的性能。结果显示,对于 Chula-AMD 数据集上的三类 AMD 分类(正常 AMD vs. 干性 AMD vs. 湿性 AMD),我们训练模型的平均准确率、精确度、灵敏度和特异性分别约为 93.40%、92.15%、91.27% 和 96.57%。在独立数据集的结果测试中,训练模型的平均准确率、精确度、灵敏度和特异性分别约为 74%、20%、75.35%、74.13% 和 87.07%。与基于 CNN 的基线模型(DenseNet201)的结果相比,训练后的 ViT-AMD 模型有明显的优越性。总之,ViT-AMD 模型证明了其在协助眼科医生诊断 AMD 疾病方面的实用性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

摘要图片

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
ViT-AMD: A New Deep Learning Model for Age-Related Macular Degeneration Diagnosis From Fundus Images

Age-related macular degeneration (AMD) diagnosis using fundus images is one of the critical missions of the eye-care screening program in many countries. Various proposed deep learning models have been studied for this research interest, which aim to achieve the mission and outperform human-based approaches. However, research efforts are still required for the improvement of model classification accuracy, sensitivity, and specificity values. In this study, we proposed the model named as ViT-AMD, which is based on the latest Vision Transformer (ViT) structure, to diagnosis a fundus image as normal, dry AMD, or wet AMD types. Unlike convolution neural network models, ViT consists of the attention map layers, which show more effective performance for image classification task. Our training process is based on the 5-fold cross-validation and transfer learning techniques using Chula-AMD dataset at the Department of Ophthalmology, the King Chulalongkorn Memorial Hospital, Bangkok. Furthermore, we also test the performance of trained model using an independent image datasets. The results showed that for the 3-classes AMD classification (normal vs. dry AMD vs. wet AMD) on the Chula-AMD dataset, the averaged accuracy, precision, sensitivity, and specificity of our trained model are about 93.40%, 92.15%, 91.27%, and 96.57%, respectively. For result testing on independent datasets, the averaged accuracy, precision, sensitivity, and specificity of trained model are about 74, 20%, 75.35%, 74.13%, and 87.07%, respectively. Compared with the results from the baseline CNN-based model (DenseNet201), the trained ViT-AMD model has outperformed significantly. In conclusion, the ViT-AMD model have proved their usefulness to assist the ophthalmologist to diagnosis the AMD disease.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
International Journal of Intelligent Systems
International Journal of Intelligent Systems 工程技术-计算机:人工智能
CiteScore
11.30
自引率
14.30%
发文量
304
审稿时长
9 months
期刊介绍: The International Journal of Intelligent Systems serves as a forum for individuals interested in tapping into the vast theories based on intelligent systems construction. With its peer-reviewed format, the journal explores several fascinating editorials written by today''s experts in the field. Because new developments are being introduced each day, there''s much to be learned — examination, analysis creation, information retrieval, man–computer interactions, and more. The International Journal of Intelligent Systems uses charts and illustrations to demonstrate these ground-breaking issues, and encourages readers to share their thoughts and experiences.
期刊最新文献
K-Means Centroids Initialization Based on Differentiation Between Instances Attributes ViT-AMD: A New Deep Learning Model for Age-Related Macular Degeneration Diagnosis From Fundus Images Switched Observer-Based Event-Triggered Safety Control for Delayed Networked Control Systems Under Aperiodic Cyber attacks An Innovative Application of Swarm-Based Algorithms for Peer Clustering Deepfake Detection Based on the Adaptive Fusion of Spatial-Frequency Features
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1