RetOCTNet: Deep Learning-Based Segmentation of OCT Images Following Retinal Ganglion Cell Injury.

IF 2.6 3区 医学 Q2 OPHTHALMOLOGY Translational Vision Science & Technology Pub Date : 2025-02-03 DOI:10.1167/tvst.14.2.4
Gabriela Sanchez-Rodriguez, Linjiang Lou, Machelle T Pardue, Andrew J Feola
{"title":"RetOCTNet: Deep Learning-Based Segmentation of OCT Images Following Retinal Ganglion Cell Injury.","authors":"Gabriela Sanchez-Rodriguez, Linjiang Lou, Machelle T Pardue, Andrew J Feola","doi":"10.1167/tvst.14.2.4","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>We present RetOCTNet, a deep learning tool to segment the retinal nerve fiber layer (RNFL) and total retinal thickness automatically from optical coherence tomography (OCT) scans in rats following retinal ganglion cell (RGC) injury.</p><p><strong>Methods: </strong>We created unilateral RGC injury by ocular hypertension (OHT) or optic nerve crush (ONC), and contralateral eyes were not injured. We manually segmented the RNFL and total retina of 3.0-mm radial OCT scans (1000 A-scans per B-scan, 20 frames per B-scan) as ground truth (n = 192 scans). We used these segmentations for training (80%), testing (10%), and validation (10%) to optimize the F1 score. To determine the generalizability of RetOCTNet, we tested it on volumetric scans of a separate cohort at baseline and 4, 8, and 12 weeks post-ONC.</p><p><strong>Results: </strong>RetOCTNet's segmentations achieved high F1 scores relative to the ground-truth segmentations created by human annotators: 0.88 (RNFL) and 0.98 (retinal thickness) for control eyes, 0.84 and 0.98 for OHT eyes, and 0.78 and 0.96 for ONC eyes, respectively. On volumetric scans 12 weeks post-ONC, RetOCTNet calculated thinning of 29.49% and 10.82% in the RNFL and retina at the optic nerve head (ONH) and thinning of 38.34% and 9.85% in the RNFL and retina superior to the ONH.</p><p><strong>Conclusions: </strong>RetOCTNet can segment the RNFL and total retinal thickness of both radial and volume OCT scans. RetOCTNet can be applied to longitudinally monitor RNFL in rodent models of RGC injury.</p><p><strong>Translational relevance: </strong>This tool automates RNFL and retinal thickness measurement for rat OCT scans following RGC injury, reducing analysis time and increasing the consistency between studies.</p>","PeriodicalId":23322,"journal":{"name":"Translational Vision Science & Technology","volume":"14 2","pages":"4"},"PeriodicalIF":2.6000,"publicationDate":"2025-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11801391/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Translational Vision Science & Technology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1167/tvst.14.2.4","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose: We present RetOCTNet, a deep learning tool to segment the retinal nerve fiber layer (RNFL) and total retinal thickness automatically from optical coherence tomography (OCT) scans in rats following retinal ganglion cell (RGC) injury.

Methods: We created unilateral RGC injury by ocular hypertension (OHT) or optic nerve crush (ONC), and contralateral eyes were not injured. We manually segmented the RNFL and total retina of 3.0-mm radial OCT scans (1000 A-scans per B-scan, 20 frames per B-scan) as ground truth (n = 192 scans). We used these segmentations for training (80%), testing (10%), and validation (10%) to optimize the F1 score. To determine the generalizability of RetOCTNet, we tested it on volumetric scans of a separate cohort at baseline and 4, 8, and 12 weeks post-ONC.

Results: RetOCTNet's segmentations achieved high F1 scores relative to the ground-truth segmentations created by human annotators: 0.88 (RNFL) and 0.98 (retinal thickness) for control eyes, 0.84 and 0.98 for OHT eyes, and 0.78 and 0.96 for ONC eyes, respectively. On volumetric scans 12 weeks post-ONC, RetOCTNet calculated thinning of 29.49% and 10.82% in the RNFL and retina at the optic nerve head (ONH) and thinning of 38.34% and 9.85% in the RNFL and retina superior to the ONH.

Conclusions: RetOCTNet can segment the RNFL and total retinal thickness of both radial and volume OCT scans. RetOCTNet can be applied to longitudinally monitor RNFL in rodent models of RGC injury.

Translational relevance: This tool automates RNFL and retinal thickness measurement for rat OCT scans following RGC injury, reducing analysis time and increasing the consistency between studies.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
RetOCTNet:基于深度学习的视网膜神经节细胞损伤OCT图像分割
目的:我们提出了一种深度学习工具RetOCTNet,用于从视网膜神经节细胞(RGC)损伤大鼠的光学相干断层扫描(OCT)中自动分割视网膜神经纤维层(RNFL)和视网膜总厚度。方法:以高眼压(OHT)或视神经压迫(ONC)造成单侧RGC损伤,对侧眼无损伤。我们手动分割3.0 mm径向OCT扫描的RNFL和总视网膜(每次b扫描1000次a扫描,每次b扫描20帧)作为基础真值(n = 192次扫描)。我们将这些分割用于训练(80%)、测试(10%)和验证(10%),以优化F1分数。为了确定RetOCTNet的普遍性,我们在基线和onc后4、8和12周的单独队列中进行了体积扫描测试。结果:相对于人类注释者创建的ground-truth分割,RetOCTNet的分割获得了很高的F1分数:对照眼的RNFL和视网膜厚度分别为0.88和0.98,OHT眼的RNFL和视网膜厚度分别为0.84和0.98,ONC眼的RNFL和视网膜厚度分别为0.78和0.96。在onc后12周的体积扫描中,RetOCTNet计算视神经头(ONH)的RNFL和视网膜变薄29.49%和10.82%,RNFL和视网膜变薄38.34%和9.85%优于ONH。结论:RetOCTNet可以分割径向和体积OCT扫描的RNFL和视网膜总厚度。RetOCTNet可用于RGC损伤啮齿动物模型的RNFL纵向监测。翻译相关性:该工具可自动测量RGC损伤后大鼠OCT扫描的RNFL和视网膜厚度,减少分析时间并增加研究之间的一致性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Translational Vision Science & Technology
Translational Vision Science & Technology Engineering-Biomedical Engineering
CiteScore
5.70
自引率
3.30%
发文量
346
审稿时长
25 weeks
期刊介绍: Translational Vision Science & Technology (TVST), an official journal of the Association for Research in Vision and Ophthalmology (ARVO), an international organization whose purpose is to advance research worldwide into understanding the visual system and preventing, treating and curing its disorders, is an online, open access, peer-reviewed journal emphasizing multidisciplinary research that bridges the gap between basic research and clinical care. A highly qualified and diverse group of Associate Editors and Editorial Board Members is led by Editor-in-Chief Marco Zarbin, MD, PhD, FARVO. The journal covers a broad spectrum of work, including but not limited to: Applications of stem cell technology for regenerative medicine, Development of new animal models of human diseases, Tissue bioengineering, Chemical engineering to improve virus-based gene delivery, Nanotechnology for drug delivery, Design and synthesis of artificial extracellular matrices, Development of a true microsurgical operating environment, Refining data analysis algorithms to improve in vivo imaging technology, Results of Phase 1 clinical trials, Reverse translational ("bedside to bench") research. TVST seeks manuscripts from scientists and clinicians with diverse backgrounds ranging from basic chemistry to ophthalmic surgery that will advance or change the way we understand and/or treat vision-threatening diseases. TVST encourages the use of color, multimedia, hyperlinks, program code and other digital enhancements.
期刊最新文献
Diagnosis of High Intracranial Pressure by Non-Optic Nerve Retinal Image Features. Short-Term Effects of Orthokeratology Treatment Zone Decentration on Choroidal Thickness and Vasculature. Targeting the Hedgehog Signaling Pathway in the Retina Using Quercetin-Loaded Lipid Nanoparticles for Myopia Control. OptiDON-Guard: An Ensemble Model for Identification of Dysthyroid Optic Neuropathy Using Optical Coherence Tomography Angiography. An Evaluation of the Mean Teacher Framework for Semi-Supervised Cataract Surgical Image Segmentation.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1