Ariana M Familiar, Neda Khalili, Nastaran Khalili, Cassidy Schuman, Evan Grove, Karthik Viswanathan, Jakob Seidlitz, Aaron Alexander-Bloch, Anna Zapaishchykova, Benjamin H Kann, Arastoo Vossough, Phillip B Storm, Adam C Resnick, Anahita Fathi Kazerooni, Ali Nabavizadeh
{"title":"增强神经科学数据共享的能力:小儿脑部核磁共振成像的深度学习去身份化方法。","authors":"Ariana M Familiar, Neda Khalili, Nastaran Khalili, Cassidy Schuman, Evan Grove, Karthik Viswanathan, Jakob Seidlitz, Aaron Alexander-Bloch, Anna Zapaishchykova, Benjamin H Kann, Arastoo Vossough, Phillip B Storm, Adam C Resnick, Anahita Fathi Kazerooni, Ali Nabavizadeh","doi":"10.3174/ajnr.A8581","DOIUrl":null,"url":null,"abstract":"<p><strong>Background and purpose: </strong>Privacy concerns, such as identifiable facial features within brain scans, have hindered the availability of pediatric neuroimaging datasets for research. Consequently, pediatric neuroscience research lags adult counterparts, particularly in rare disease and under-represented populations. The removal of face regions (image defacing) can mitigate this, however existing defacing tools often fail with pediatric cases and diverse image types, leaving a critical gap in data accessibility. Given recent NIH data sharing mandates, novel solutions are a critical need.</p><p><strong>Materials and methods: </strong>To develop an AI-powered tool for automatic defacing of pediatric brain MRIs, deep learning methodologies (nnU-Net) were employed using a large, diverse multi-institutional dataset of clinical radiology images. This included multi-parametric MRIs (T1w, T1w-contrast enhanced, T2w, T2w-FLAIR) with 976 total images from 208 brain tumor patients (Children's Brain Tumor Network, CBTN) and 36 clinical control patients (Scans with Limited Imaging Pathology, SLIP) ranging in age from 7 days to 21 years old.</p><p><strong>Results: </strong>Face and ear removal accuracy for withheld testing data was the primary measure of model performance. Potential influences of defacing on downstream research usage were evaluated with standard image processing and AI-based pipelines. Group-level statistical trends were compared between original (non-defaced) and defaced images. Across image types, the model had high accuracy for removing face regions (mean accuracy, 98%; <i>N</i>=98 subjects/392 images), with lower performance for removal of ears (73%). Analysis of global and regional brain measures (SLIP cohort) showed minimal differences between original and defaced outputs (mean <i>r</i> <sub>S</sub>=0.93, all <i>p</i> < 0.0001). AI-generated whole brain and tumor volumes (CBTN cohort) and temporalis muscle metrics (volume, cross-sectional area, centile scores; SLIP cohort) were not significantly affected by image defacing (all <i>r</i> <sub>S</sub>>0.9, <i>p</i><0.0001).</p><p><strong>Conclusions: </strong>The defacing model demonstrates efficacy in removing facial regions across multiple MRI types and exhibits minimal impact on downstream research usage. A software package with the trained model is freely provided for wider use and further development (pediatric-auto-defacer; https://github.com/d3b-center/pediatric-auto-defacer-public). By offering a solution tailored to pediatric cases and multiple MRI sequences, this defacing tool will expedite research efforts and promote broader adoption of data sharing practices within the neuroscience community.</p><p><strong>Abbreviations: </strong>AI = artificial intelligence; CBTN = Children's Brain Tumor Network; CSA = cross-sectional area; SLIP = Scans with Limited Imaging Pathology; TMT = temporalis muscle thickness; NIH = National Institute of Health; LH = left hemisphere; RH = right hemisphere.</p>","PeriodicalId":93863,"journal":{"name":"AJNR. American journal of neuroradiology","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Empowering Data Sharing in Neuroscience: A Deep Learning De-identification Method for Pediatric Brain MRIs.\",\"authors\":\"Ariana M Familiar, Neda Khalili, Nastaran Khalili, Cassidy Schuman, Evan Grove, Karthik Viswanathan, Jakob Seidlitz, Aaron Alexander-Bloch, Anna Zapaishchykova, Benjamin H Kann, Arastoo Vossough, Phillip B Storm, Adam C Resnick, Anahita Fathi Kazerooni, Ali Nabavizadeh\",\"doi\":\"10.3174/ajnr.A8581\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background and purpose: </strong>Privacy concerns, such as identifiable facial features within brain scans, have hindered the availability of pediatric neuroimaging datasets for research. Consequently, pediatric neuroscience research lags adult counterparts, particularly in rare disease and under-represented populations. The removal of face regions (image defacing) can mitigate this, however existing defacing tools often fail with pediatric cases and diverse image types, leaving a critical gap in data accessibility. Given recent NIH data sharing mandates, novel solutions are a critical need.</p><p><strong>Materials and methods: </strong>To develop an AI-powered tool for automatic defacing of pediatric brain MRIs, deep learning methodologies (nnU-Net) were employed using a large, diverse multi-institutional dataset of clinical radiology images. This included multi-parametric MRIs (T1w, T1w-contrast enhanced, T2w, T2w-FLAIR) with 976 total images from 208 brain tumor patients (Children's Brain Tumor Network, CBTN) and 36 clinical control patients (Scans with Limited Imaging Pathology, SLIP) ranging in age from 7 days to 21 years old.</p><p><strong>Results: </strong>Face and ear removal accuracy for withheld testing data was the primary measure of model performance. Potential influences of defacing on downstream research usage were evaluated with standard image processing and AI-based pipelines. Group-level statistical trends were compared between original (non-defaced) and defaced images. Across image types, the model had high accuracy for removing face regions (mean accuracy, 98%; <i>N</i>=98 subjects/392 images), with lower performance for removal of ears (73%). Analysis of global and regional brain measures (SLIP cohort) showed minimal differences between original and defaced outputs (mean <i>r</i> <sub>S</sub>=0.93, all <i>p</i> < 0.0001). AI-generated whole brain and tumor volumes (CBTN cohort) and temporalis muscle metrics (volume, cross-sectional area, centile scores; SLIP cohort) were not significantly affected by image defacing (all <i>r</i> <sub>S</sub>>0.9, <i>p</i><0.0001).</p><p><strong>Conclusions: </strong>The defacing model demonstrates efficacy in removing facial regions across multiple MRI types and exhibits minimal impact on downstream research usage. A software package with the trained model is freely provided for wider use and further development (pediatric-auto-defacer; https://github.com/d3b-center/pediatric-auto-defacer-public). By offering a solution tailored to pediatric cases and multiple MRI sequences, this defacing tool will expedite research efforts and promote broader adoption of data sharing practices within the neuroscience community.</p><p><strong>Abbreviations: </strong>AI = artificial intelligence; CBTN = Children's Brain Tumor Network; CSA = cross-sectional area; SLIP = Scans with Limited Imaging Pathology; TMT = temporalis muscle thickness; NIH = National Institute of Health; LH = left hemisphere; RH = right hemisphere.</p>\",\"PeriodicalId\":93863,\"journal\":{\"name\":\"AJNR. American journal of neuroradiology\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-11-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"AJNR. American journal of neuroradiology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3174/ajnr.A8581\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"AJNR. American journal of neuroradiology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3174/ajnr.A8581","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
背景和目的:隐私问题(如脑部扫描中可识别的面部特征)阻碍了儿科神经成像数据集的研究。因此,儿科神经科学研究落后于成人研究,尤其是在罕见疾病和代表性不足的人群方面。去除面部区域(图像篡改)可以缓解这一问题,但现有的篡改工具往往无法处理儿科病例和不同的图像类型,从而在数据可访问性方面留下了关键的空白。鉴于最近美国国立卫生研究院(NIH)的数据共享规定,新型解决方案是一项关键需求:为了开发一种人工智能驱动的小儿脑部核磁共振成像自动去污工具,我们采用了深度学习方法(nnU-Net),使用了一个大型、多样化的多机构临床放射学图像数据集。该数据集包括多参数核磁共振成像(T1w、T1w-对比增强、T2w、T2w-FLAIR),共有 976 张图像,分别来自 208 名脑肿瘤患者(儿童脑肿瘤网络,CBTN)和 36 名临床对照患者(有限影像病理学扫描,SLIP),年龄从 7 天到 21 岁不等:对扣留的测试数据进行面部和耳朵移除的准确性是衡量模型性能的主要标准。使用标准图像处理和基于人工智能的管道评估了玷污对下游研究使用的潜在影响。对原始图像(未玷污)和玷污图像进行了组级统计趋势比较。在所有图像类型中,该模型去除面部区域的准确率较高(平均准确率为 98%;受试者人数=98 人/392 张图像),而去除耳朵的准确率较低(73%)。全局和区域大脑测量分析(SLIP 队列)显示,原始输出和污损输出之间的差异极小(平均 r S=0.93,所有 p <0.0001)。人工智能生成的全脑和肿瘤体积(CBTN 队列)以及颞肌指标(体积、横截面积、百分位数;SLIP 队列)没有受到图像去污的显著影响(所有 r S 均大于 0.9,p 结论:去污模型在多种核磁共振成像类型中都能有效去除面部区域,而且对下游研究使用的影响极小。我们免费提供了一个包含训练有素模型的软件包,供更广泛使用和进一步开发(phiatric-auto-defacer; https://github.com/d3b-center/pediatric-auto-defacer-public)。通过提供针对儿科病例和多种核磁共振成像序列的解决方案,该去污工具将加快研究工作,并促进神经科学界更广泛地采用数据共享做法:缩写:AI=人工智能;CBTN=儿童脑肿瘤网络;CSA=横截面积;SLIP=有限病理成像扫描;TMT=颞肌厚度;NIH=美国国立卫生研究院;LH=左半球;RH=右半球。
Empowering Data Sharing in Neuroscience: A Deep Learning De-identification Method for Pediatric Brain MRIs.
Background and purpose: Privacy concerns, such as identifiable facial features within brain scans, have hindered the availability of pediatric neuroimaging datasets for research. Consequently, pediatric neuroscience research lags adult counterparts, particularly in rare disease and under-represented populations. The removal of face regions (image defacing) can mitigate this, however existing defacing tools often fail with pediatric cases and diverse image types, leaving a critical gap in data accessibility. Given recent NIH data sharing mandates, novel solutions are a critical need.
Materials and methods: To develop an AI-powered tool for automatic defacing of pediatric brain MRIs, deep learning methodologies (nnU-Net) were employed using a large, diverse multi-institutional dataset of clinical radiology images. This included multi-parametric MRIs (T1w, T1w-contrast enhanced, T2w, T2w-FLAIR) with 976 total images from 208 brain tumor patients (Children's Brain Tumor Network, CBTN) and 36 clinical control patients (Scans with Limited Imaging Pathology, SLIP) ranging in age from 7 days to 21 years old.
Results: Face and ear removal accuracy for withheld testing data was the primary measure of model performance. Potential influences of defacing on downstream research usage were evaluated with standard image processing and AI-based pipelines. Group-level statistical trends were compared between original (non-defaced) and defaced images. Across image types, the model had high accuracy for removing face regions (mean accuracy, 98%; N=98 subjects/392 images), with lower performance for removal of ears (73%). Analysis of global and regional brain measures (SLIP cohort) showed minimal differences between original and defaced outputs (mean rS=0.93, all p < 0.0001). AI-generated whole brain and tumor volumes (CBTN cohort) and temporalis muscle metrics (volume, cross-sectional area, centile scores; SLIP cohort) were not significantly affected by image defacing (all rS>0.9, p<0.0001).
Conclusions: The defacing model demonstrates efficacy in removing facial regions across multiple MRI types and exhibits minimal impact on downstream research usage. A software package with the trained model is freely provided for wider use and further development (pediatric-auto-defacer; https://github.com/d3b-center/pediatric-auto-defacer-public). By offering a solution tailored to pediatric cases and multiple MRI sequences, this defacing tool will expedite research efforts and promote broader adoption of data sharing practices within the neuroscience community.
Abbreviations: AI = artificial intelligence; CBTN = Children's Brain Tumor Network; CSA = cross-sectional area; SLIP = Scans with Limited Imaging Pathology; TMT = temporalis muscle thickness; NIH = National Institute of Health; LH = left hemisphere; RH = right hemisphere.