首页 > 最新文献

Proceedings of the ACM Conference on Health, Inference, and Learning最新文献

英文 中文
CheXternal: generalization of deep learning models for chest X-ray interpretation to photos of chest X-rays and external clinical settings CheXternal:将胸部x线解读的深度学习模型推广到胸部x线照片和外部临床设置
Pub Date : 2021-02-17 DOI: 10.1145/3450439.3451876
P. Rajpurkar, Anirudh Joshi, A. Pareek, A. Ng, M. Lungren
Recent advances in training deep learning models have demonstrated the potential to provide accurate chest X-ray interpretation and increase access to radiology expertise. However, poor generalization due to data distribution shifts in clinical settings is a key barrier to implementation. In this study, we measured the diagnostic performance for 8 different chest X-ray models when applied to (1) smartphone photos of chest X-rays and (2) external datasets without any finetuning. All models were developed by different groups and submitted to the CheXpert challenge, and re-applied to test datasets without further tuning. We found that (1) on photos of chest X-rays, all 8 models experienced a statistically significant drop in task performance, but only 3 performed significantly worse than radiologists on average, and (2) on the external set, none of the models performed statistically significantly worse than radiologists, and five models performed statistically significantly better than radiologists. Our results demonstrate that some chest X-ray models, under clinically relevant distribution shifts, were comparable to radiologists while other models were not. Future work should investigate aspects of model training procedures and dataset collection that influence generalization in the presence of data distribution shifts.
训练深度学习模型的最新进展已经证明了提供准确的胸部x射线解释和增加获得放射学专业知识的潜力。然而,由于临床环境中数据分布的变化而导致的泛化不良是实施的主要障碍。在这项研究中,我们测量了8种不同的胸部x线模型在应用于(1)智能手机的胸部x线照片和(2)没有任何微调的外部数据集时的诊断性能。所有模型都由不同的小组开发,并提交给CheXpert挑战,并在没有进一步调优的情况下重新应用于测试数据集。我们发现(1)在胸部x光照片上,所有8个模型的任务表现都有统计学上的显著下降,但只有3个模型的平均表现明显低于放射科医生;(2)在外部集合上,没有一个模型的任务表现在统计学上显著低于放射科医生,有5个模型的任务表现在统计学上显著优于放射科医生。我们的研究结果表明,在临床相关的分布变化下,一些胸部x线模型与放射科医生具有可比性,而其他模型则不然。未来的工作应该研究在数据分布变化的情况下影响泛化的模型训练程序和数据集收集方面。
{"title":"CheXternal: generalization of deep learning models for chest X-ray interpretation to photos of chest X-rays and external clinical settings","authors":"P. Rajpurkar, Anirudh Joshi, A. Pareek, A. Ng, M. Lungren","doi":"10.1145/3450439.3451876","DOIUrl":"https://doi.org/10.1145/3450439.3451876","url":null,"abstract":"Recent advances in training deep learning models have demonstrated the potential to provide accurate chest X-ray interpretation and increase access to radiology expertise. However, poor generalization due to data distribution shifts in clinical settings is a key barrier to implementation. In this study, we measured the diagnostic performance for 8 different chest X-ray models when applied to (1) smartphone photos of chest X-rays and (2) external datasets without any finetuning. All models were developed by different groups and submitted to the CheXpert challenge, and re-applied to test datasets without further tuning. We found that (1) on photos of chest X-rays, all 8 models experienced a statistically significant drop in task performance, but only 3 performed significantly worse than radiologists on average, and (2) on the external set, none of the models performed statistically significantly worse than radiologists, and five models performed statistically significantly better than radiologists. Our results demonstrate that some chest X-ray models, under clinically relevant distribution shifts, were comparable to radiologists while other models were not. Future work should investigate aspects of model training procedures and dataset collection that influence generalization in the presence of data distribution shifts.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-02-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81221608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
CheXtransfer: performance and parameter efficiency of ImageNet models for chest X-Ray interpretation CheXtransfer: ImageNet模型用于胸部x线解译的性能和参数效率
Pub Date : 2021-01-18 DOI: 10.1145/3450439.3451867
Alexander Ke, William Ellsworth, O. Banerjee, A. Ng, P. Rajpurkar
Deep learning methods for chest X-ray interpretation typically rely on pretrained models developed for ImageNet. This paradigm assumes that better ImageNet architectures perform better on chest X-ray tasks and that ImageNet-pretrained weights provide a performance boost over random initialization. In this work, we compare the transfer performance and parameter efficiency of 16 popular convolutional architectures on a large chest X-ray dataset (CheXpert) to investigate these assumptions. First, we find no relationship between ImageNet performance and CheXpert performance for both models without pretraining and models with pretraining. Second, we find that, for models without pretraining, the choice of model family influences performance more than size within a family for medical imaging tasks. Third, we observe that ImageNet pretraining yields a statistically significant boost in performance across architectures, with a higher boost for smaller architectures. Fourth, we examine whether ImageNet architectures are unnecessarily large for CheXpert by truncating final blocks from pretrained models, and find that we can make models 3.25x more parameter-efficient on average without a statistically significant drop in performance. Our work contributes new experimental evidence about the relation of ImageNet to chest x-ray interpretation performance.
胸部x射线解释的深度学习方法通常依赖于为ImageNet开发的预训练模型。这个范例假设更好的ImageNet架构在胸部x射线任务上表现更好,并且ImageNet预训练的权重比随机初始化提供了性能提升。在这项工作中,我们比较了16种流行的卷积架构在大型胸部x射线数据集(CheXpert)上的传输性能和参数效率,以研究这些假设。首先,我们发现没有预训练的模型和有预训练的模型的ImageNet性能和CheXpert性能之间没有关系。其次,我们发现,对于没有预训练的模型,模型族的选择对医学成像任务的表现的影响大于家族的大小。第三,我们观察到ImageNet预训练在跨架构的性能上产生了统计上显著的提升,对于较小的架构有更高的提升。第四,我们通过截断预训练模型的最终块来检查ImageNet架构对于CheXpert是否不必要地大,并发现我们可以使模型的参数效率平均提高3.25倍,而性能却没有统计学上的显著下降。我们的工作为ImageNet与胸部x射线解释性能的关系提供了新的实验证据。
{"title":"CheXtransfer: performance and parameter efficiency of ImageNet models for chest X-Ray interpretation","authors":"Alexander Ke, William Ellsworth, O. Banerjee, A. Ng, P. Rajpurkar","doi":"10.1145/3450439.3451867","DOIUrl":"https://doi.org/10.1145/3450439.3451867","url":null,"abstract":"Deep learning methods for chest X-ray interpretation typically rely on pretrained models developed for ImageNet. This paradigm assumes that better ImageNet architectures perform better on chest X-ray tasks and that ImageNet-pretrained weights provide a performance boost over random initialization. In this work, we compare the transfer performance and parameter efficiency of 16 popular convolutional architectures on a large chest X-ray dataset (CheXpert) to investigate these assumptions. First, we find no relationship between ImageNet performance and CheXpert performance for both models without pretraining and models with pretraining. Second, we find that, for models without pretraining, the choice of model family influences performance more than size within a family for medical imaging tasks. Third, we observe that ImageNet pretraining yields a statistically significant boost in performance across architectures, with a higher boost for smaller architectures. Fourth, we examine whether ImageNet architectures are unnecessarily large for CheXpert by truncating final blocks from pretrained models, and find that we can make models 3.25x more parameter-efficient on average without a statistically significant drop in performance. Our work contributes new experimental evidence about the relation of ImageNet to chest x-ray interpretation performance.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73858761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 81
ACM CHIL '21: ACM Conference on Health, Inference, and Learning, Virtual Event, USA, April 8-9, 2021 ACM CHIL '21: ACM健康、推理和学习会议,虚拟事件,美国,2021年4月8日至9日
{"title":"ACM CHIL '21: ACM Conference on Health, Inference, and Learning, Virtual Event, USA, April 8-9, 2021","authors":"","doi":"10.1145/3450439","DOIUrl":"https://doi.org/10.1145/3450439","url":null,"abstract":"","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75547599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
iGOS++: integrated gradient optimized saliency by bilateral perturbations igos++:双侧扰动下的积分梯度优化显著性
Pub Date : 2021-01-01 DOI: 10.1145/3450439.3451865
S. Khorram, T. Lawson, Fuxin Li
The black-box nature of the deep networks makes the explanation for "why" they make certain predictions extremely challenging. Saliency maps are one of the most widely-used local explanation tools to alleviate this problem. One of the primary approaches for generating saliency maps is by optimizing for a mask over the input dimensions so that the output of the network for a given class is influenced the most. However, prior work only studies such influence by removing evidence from the input. In this paper, we present iGOS++, a framework to generate saliency maps for blackbox networks by considering both removal and preservation of evidence. Additionally, we introduce the bilateral total variation term to the optimization that improves the continuity of the saliency map especially under high resolution and with thin object parts. We validate the capabilities of iGOS++ by extensive experiments and comparison against state-of-the-art saliency map methods. Our results show significant improvement in locating salient regions that are directly interpretable by humans. Besides, we showcased the capabilities of our method, iGOS++, in a real-world application of AI on medical data: the task of classifying COVID-19 cases from x-ray images. To our surprise, we discovered that sometimes the classifier is overfitted to the text characters printed on the x-ray images when performing classification rather than focusing on the evidence in the lungs. Fixing this overfitting issue by data cleansing significantly improved the precision and recall of the classifier.
深度网络的黑箱特性使得解释它们“为什么”做出某些预测极具挑战性。显著性图是缓解这一问题的最广泛使用的局部解释工具之一。生成显著性图的主要方法之一是通过优化输入维度上的掩码,以便对给定类的网络输出影响最大。然而,先前的工作只是通过从输入中删除证据来研究这种影响。在本文中,我们提出了igos++,一个框架来生成显著性地图的黑箱网络,同时考虑移除和保留证据。此外,我们在优化中引入了双边总变分项,提高了显著性图的连续性,特别是在高分辨率和薄目标部分的情况下。我们通过大量的实验和与最先进的显著性图方法的比较来验证igos++的功能。我们的研究结果表明,在定位人类可直接解释的显著区域方面有了显著的改进。此外,我们还展示了我们的方法igos++在人工智能在医疗数据上的实际应用中的功能:从x射线图像中分类COVID-19病例的任务。令我们惊讶的是,我们发现有时分类器在进行分类时过度拟合x射线图像上的文本字符,而不是专注于肺部的证据。通过数据清理修复这个过拟合问题显著提高了分类器的精度和召回率。
{"title":"iGOS++: integrated gradient optimized saliency by bilateral perturbations","authors":"S. Khorram, T. Lawson, Fuxin Li","doi":"10.1145/3450439.3451865","DOIUrl":"https://doi.org/10.1145/3450439.3451865","url":null,"abstract":"The black-box nature of the deep networks makes the explanation for \"why\" they make certain predictions extremely challenging. Saliency maps are one of the most widely-used local explanation tools to alleviate this problem. One of the primary approaches for generating saliency maps is by optimizing for a mask over the input dimensions so that the output of the network for a given class is influenced the most. However, prior work only studies such influence by removing evidence from the input. In this paper, we present iGOS++, a framework to generate saliency maps for blackbox networks by considering both removal and preservation of evidence. Additionally, we introduce the bilateral total variation term to the optimization that improves the continuity of the saliency map especially under high resolution and with thin object parts. We validate the capabilities of iGOS++ by extensive experiments and comparison against state-of-the-art saliency map methods. Our results show significant improvement in locating salient regions that are directly interpretable by humans. Besides, we showcased the capabilities of our method, iGOS++, in a real-world application of AI on medical data: the task of classifying COVID-19 cases from x-ray images. To our surprise, we discovered that sometimes the classifier is overfitted to the text characters printed on the x-ray images when performing classification rather than focusing on the evidence in the lungs. Fixing this overfitting issue by data cleansing significantly improved the precision and recall of the classifier.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81277946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Concept-based model explanations for electronic health records 基于概念的电子健康记录模型解释
Pub Date : 2020-12-03 DOI: 10.1145/3450439.3451858
Sebastien Baur, Shaobo Hou, Eric Loreaux, Diana Mincu, A. Mottram, Ivan V. Protsyuk, Nenad Tomašev, Martin G. Seneviratne, Alan Karthikesanlingam, J. Schrouff
Recurrent Neural Networks (RNNs) are often used for sequential modeling of adverse outcomes in electronic health records (EHRs) due to their ability to encode past clinical states. These deep, recurrent architectures have displayed increased performance compared to other modeling approaches in a number of tasks, fueling the interest in deploying deep models in clinical settings. One of the key elements in ensuring safe model deployment and building user trust is model explainability. Testing with Concept Activation Vectors (TCAV) has recently been introduced as a way of providing human-understandable explanations by comparing high-level concepts to the network's gradients. While the technique has shown promising results in real-world imaging applications, it has not been applied to structured temporal inputs. To enable an application of TCAV to sequential predictions in the EHR, we propose an extension of the method to time series data. We evaluate the proposed approach on an open EHR benchmark from the intensive care unit, as well as synthetic data where we are able to better isolate individual effects.
递归神经网络(rnn)通常用于电子健康记录(EHRs)中不良后果的顺序建模,因为它们能够编码过去的临床状态。与其他建模方法相比,这些深度的、循环的架构在许多任务中显示出更高的性能,激发了在临床环境中部署深度模型的兴趣。确保安全模型部署和建立用户信任的关键因素之一是模型的可解释性。概念激活向量测试(TCAV)最近被引入,作为一种通过比较高级概念和网络梯度来提供人类可理解的解释的方法。虽然该技术在现实世界的成像应用中显示出有希望的结果,但它还没有应用于结构化的时间输入。为了使TCAV应用于EHR中的序列预测,我们提出了将该方法扩展到时间序列数据的方法。我们在重症监护病房的开放式电子病历基准以及能够更好地隔离个体影响的合成数据上评估所提出的方法。
{"title":"Concept-based model explanations for electronic health records","authors":"Sebastien Baur, Shaobo Hou, Eric Loreaux, Diana Mincu, A. Mottram, Ivan V. Protsyuk, Nenad Tomašev, Martin G. Seneviratne, Alan Karthikesanlingam, J. Schrouff","doi":"10.1145/3450439.3451858","DOIUrl":"https://doi.org/10.1145/3450439.3451858","url":null,"abstract":"Recurrent Neural Networks (RNNs) are often used for sequential modeling of adverse outcomes in electronic health records (EHRs) due to their ability to encode past clinical states. These deep, recurrent architectures have displayed increased performance compared to other modeling approaches in a number of tasks, fueling the interest in deploying deep models in clinical settings. One of the key elements in ensuring safe model deployment and building user trust is model explainability. Testing with Concept Activation Vectors (TCAV) has recently been introduced as a way of providing human-understandable explanations by comparing high-level concepts to the network's gradients. While the technique has shown promising results in real-world imaging applications, it has not been applied to structured temporal inputs. To enable an application of TCAV to sequential predictions in the EHR, we propose an extension of the method to time series data. We evaluate the proposed approach on an open EHR benchmark from the intensive care unit, as well as synthetic data where we are able to better isolate individual effects.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89243679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Self-supervised transfer learning of physiological representations from free-living wearable data 来自自由生活的可穿戴数据的生理表征的自监督迁移学习
Pub Date : 2020-11-18 DOI: 10.1145/3450439.3451863
Dimitris Spathis, I. Perez-Pozuelo, S. Brage, N. Wareham, C. Mascolo
Wearable devices such as smartwatches are becoming increasingly popular tools for objectively monitoring physical activity in free-living conditions. To date, research has primarily focused on the purely supervised task of human activity recognition, demonstrating limited success in inferring high-level health outcomes from low-level signals. Here, we present a novel self-supervised representation learning method using activity and heart rate (HR) signals without semantic labels. With a deep neural network, we set HR responses as the supervisory signal for the activity data, leveraging their underlying physiological relationship. In addition, we propose a custom quantile loss function that accounts for the long-tailed HR distribution present in the general population. We evaluate our model in the largest free-living combined-sensing dataset (comprising >280k hours of wrist accelerometer & wearable ECG data). Our contributions are two-fold: i) the pre-training task creates a model that can accurately forecast HR based only on cheap activity sensors, and ii) we leverage the information captured through this task by proposing a simple method to aggregate the learnt latent representations (embeddings) from the window-level to user-level. Notably, we show that the embeddings can generalize in various downstream tasks through transfer learning with linear classifiers, capturing physiologically meaningful, personalized information. For instance, they can be used to predict variables associated with individuals' health, fitness and demographic characteristics (AUC >70), outperforming unsupervised autoencoders and common bio-markers. Overall, we propose the first multimodal self-supervised method for behavioral and physiological data with implications for large-scale health and lifestyle monitoring. Code: https://github.com/sdimi/Step2heart.
在自由生活的条件下,智能手表等可穿戴设备正成为越来越受欢迎的客观监测身体活动的工具。迄今为止,研究主要集中在人类活动识别的纯粹监督任务上,表明在从低水平信号推断高水平健康结果方面取得的成功有限。在这里,我们提出了一种新的自监督表示学习方法,使用活动和心率(HR)信号,而不使用语义标签。通过深层神经网络,我们将HR反应设置为活动数据的监督信号,利用它们之间潜在的生理关系。此外,我们提出了一个自定义的分位数损失函数,用于解释一般人群中存在的长尾HR分布。我们在最大的自由生活组合传感数据集(包括>280k小时的手腕加速度计和可穿戴ECG数据)中评估了我们的模型。我们的贡献是双重的:i)预训练任务创建了一个仅基于廉价活动传感器就能准确预测HR的模型,ii)我们利用通过该任务捕获的信息,提出了一种简单的方法,将学习到的潜在表征(嵌入)从窗口级聚合到用户级。值得注意的是,我们表明嵌入可以通过线性分类器的迁移学习来推广各种下游任务,捕获生理上有意义的个性化信息。例如,它们可用于预测与个人健康、适应性和人口特征(AUC >70)相关的变量,优于无监督自动编码器和普通生物标记。总的来说,我们提出了第一个多模式自我监督方法,用于行为和生理数据,具有大规模健康和生活方式监测的意义。代码:https://github.com/sdimi/Step2heart。
{"title":"Self-supervised transfer learning of physiological representations from free-living wearable data","authors":"Dimitris Spathis, I. Perez-Pozuelo, S. Brage, N. Wareham, C. Mascolo","doi":"10.1145/3450439.3451863","DOIUrl":"https://doi.org/10.1145/3450439.3451863","url":null,"abstract":"Wearable devices such as smartwatches are becoming increasingly popular tools for objectively monitoring physical activity in free-living conditions. To date, research has primarily focused on the purely supervised task of human activity recognition, demonstrating limited success in inferring high-level health outcomes from low-level signals. Here, we present a novel self-supervised representation learning method using activity and heart rate (HR) signals without semantic labels. With a deep neural network, we set HR responses as the supervisory signal for the activity data, leveraging their underlying physiological relationship. In addition, we propose a custom quantile loss function that accounts for the long-tailed HR distribution present in the general population. We evaluate our model in the largest free-living combined-sensing dataset (comprising >280k hours of wrist accelerometer & wearable ECG data). Our contributions are two-fold: i) the pre-training task creates a model that can accurately forecast HR based only on cheap activity sensors, and ii) we leverage the information captured through this task by proposing a simple method to aggregate the learnt latent representations (embeddings) from the window-level to user-level. Notably, we show that the embeddings can generalize in various downstream tasks through transfer learning with linear classifiers, capturing physiologically meaningful, personalized information. For instance, they can be used to predict variables associated with individuals' health, fitness and demographic characteristics (AUC >70), outperforming unsupervised autoencoders and common bio-markers. Overall, we propose the first multimodal self-supervised method for behavioral and physiological data with implications for large-scale health and lifestyle monitoring. Code: https://github.com/sdimi/Step2heart.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80007425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Phenotypical ontology driven framework for multi-task learning 多任务学习的表型本体驱动框架
Pub Date : 2020-09-04 DOI: 10.1145/3450439.3451881
Mohamed F. Ghalwash, Zijun Yao, P. Chakraborty, James Codella, D. Sow
Despite the large number of patients in Electronic Health Records (EHRs), the subset of usable data for modeling outcomes of specific phenotypes are often imbalanced and of modest size. This can be attributed to the uneven coverage of medical concepts in EHRs. We propose OMTL, an Ontology-driven Multi-Task Learning framework, that is designed to overcome such data limitations.The key contribution of our work is the effective use of knowledge from a predefined well-established medical relationship graph (ontology) to construct a novel deep learning network architecture that mirrors this ontology. This enables common representations to be shared across related phenotypes, and was found to improve the learning performance. The proposed OMTL naturally allows for multi-task learning of different phenotypes on distinct predictive tasks. These phenotypes are tied together by their semantic relationship according to the external medical ontology. Using the publicly available MIMIC-III database, we evaluate OMTL and demonstrate its efficacy on several real patient outcome predictions over state-of-the-art multi-task learning schemes. The results of evaluating the proposed approach on six experiments show improvement in the area under ROC curve by 9% and by 8% in the area under precision-recall curve.
尽管电子健康记录(EHRs)中有大量患者,但用于特定表型建模结果的可用数据子集通常是不平衡的,并且大小适中。这可归因于电子病历中医疗概念的不均匀覆盖。我们提出了OMTL,一个本体驱动的多任务学习框架,旨在克服这种数据限制。我们工作的关键贡献是有效地利用预定义的已建立的医学关系图(本体)中的知识来构建反映该本体的新型深度学习网络架构。这使得共同表征可以在相关表型之间共享,并且被发现可以提高学习性能。提出的OMTL自然允许在不同的预测任务上对不同表型的多任务学习。这些表型根据外部医学本体的语义关系联系在一起。使用公开可用的MIMIC-III数据库,我们评估了OMTL,并证明了它在最先进的多任务学习方案中对几个真实患者预后预测的有效性。6个实验结果表明,该方法的ROC曲线下面积和precision-recall曲线下面积分别提高了9%和8%。
{"title":"Phenotypical ontology driven framework for multi-task learning","authors":"Mohamed F. Ghalwash, Zijun Yao, P. Chakraborty, James Codella, D. Sow","doi":"10.1145/3450439.3451881","DOIUrl":"https://doi.org/10.1145/3450439.3451881","url":null,"abstract":"Despite the large number of patients in Electronic Health Records (EHRs), the subset of usable data for modeling outcomes of specific phenotypes are often imbalanced and of modest size. This can be attributed to the uneven coverage of medical concepts in EHRs. We propose OMTL, an Ontology-driven Multi-Task Learning framework, that is designed to overcome such data limitations.The key contribution of our work is the effective use of knowledge from a predefined well-established medical relationship graph (ontology) to construct a novel deep learning network architecture that mirrors this ontology. This enables common representations to be shared across related phenotypes, and was found to improve the learning performance. The proposed OMTL naturally allows for multi-task learning of different phenotypes on distinct predictive tasks. These phenotypes are tied together by their semantic relationship according to the external medical ontology. Using the publicly available MIMIC-III database, we evaluate OMTL and demonstrate its efficacy on several real patient outcome predictions over state-of-the-art multi-task learning schemes. The results of evaluating the proposed approach on six experiments show improvement in the area under ROC curve by 9% and by 8% in the area under precision-recall curve.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84508295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Temporal pointwise convolutional networks for length of stay prediction in the intensive care unit 用于重症监护病房住院时间预测的时间点卷积网络
Pub Date : 2020-07-18 DOI: 10.1145/3450439.3451860
Emma Rocheteau, P. Lio’, Stephanie L. Hyland
The pressure of ever-increasing patient demand and budget restrictions make hospital bed management a daily challenge for clinical staff. Most critical is the efficient allocation of resource-heavy Intensive Care Unit (ICU) beds to the patients who need life support. Central to solving this problem is knowing for how long the current set of ICU patients are likely to stay in the unit. In this work, we propose a new deep learning model based on the combination of temporal convolution and pointwise (1x1) convolution, to solve the length of stay prediction task on the eICU and MIMIC-IV critical care datasets. The model - which we refer to as Temporal Pointwise Convolution (TPC) - is specifically designed to mitigate common challenges with Electronic Health Records, such as skewness, irregular sampling and missing data. In doing so, we have achieved significant performance benefits of 18-68% (metric and dataset dependent) over the commonly used Long-Short Term Memory (LSTM) network, and the multi-head self-attention network known as the Transformer. By adding mortality prediction as a side-task, we can improve performance further still, resulting in a mean absolute deviation of 1.55 days (eICU) and 2.28 days (MIMIC-IV) on predicting remaining length of stay.
不断增加的病人需求和预算限制的压力使医院病床管理成为临床工作人员的日常挑战。最关键的是有效地将资源繁重的重症监护病房(ICU)床位分配给需要生命支持的患者。解决这一问题的核心是了解当前ICU患者可能在病房待多久。在这项工作中,我们提出了一种新的基于时间卷积和点向(1x1)卷积相结合的深度学习模型,以解决eICU和MIMIC-IV重症监护数据集的住院时间预测任务。该模型——我们称之为时间点卷积(TPC)——是专门为减轻电子健康记录的常见挑战而设计的,比如偏度、不规则采样和数据缺失。在这样做的过程中,我们已经取得了18-68%的显著性能优势(指标和数据集依赖),超过了常用的长短期记忆(LSTM)网络,以及被称为Transformer的多头自关注网络。通过增加死亡率预测作为副任务,我们可以进一步提高性能,导致预测剩余住院时间的平均绝对偏差为1.55天(eICU)和2.28天(MIMIC-IV)。
{"title":"Temporal pointwise convolutional networks for length of stay prediction in the intensive care unit","authors":"Emma Rocheteau, P. Lio’, Stephanie L. Hyland","doi":"10.1145/3450439.3451860","DOIUrl":"https://doi.org/10.1145/3450439.3451860","url":null,"abstract":"The pressure of ever-increasing patient demand and budget restrictions make hospital bed management a daily challenge for clinical staff. Most critical is the efficient allocation of resource-heavy Intensive Care Unit (ICU) beds to the patients who need life support. Central to solving this problem is knowing for how long the current set of ICU patients are likely to stay in the unit. In this work, we propose a new deep learning model based on the combination of temporal convolution and pointwise (1x1) convolution, to solve the length of stay prediction task on the eICU and MIMIC-IV critical care datasets. The model - which we refer to as Temporal Pointwise Convolution (TPC) - is specifically designed to mitigate common challenges with Electronic Health Records, such as skewness, irregular sampling and missing data. In doing so, we have achieved significant performance benefits of 18-68% (metric and dataset dependent) over the commonly used Long-Short Term Memory (LSTM) network, and the multi-head self-attention network known as the Transformer. By adding mortality prediction as a side-task, we can improve performance further still, resulting in a mean absolute deviation of 1.55 days (eICU) and 2.28 days (MIMIC-IV) on predicting remaining length of stay.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82370769","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
BMM-Net: automatic segmentation of edema in optical coherence tomography based on boundary detection and multi-scale network BMM-Net:基于边界检测和多尺度网络的光学相干层析成像水肿自动分割
Pub Date : 2020-04-02 DOI: 10.1145/3368555.3384447
Ruru Zhang, Jiawen He, Shenda Shi, E. Haihong, Zhonghong Ou, Meina Song
Retinal effusions and cysts caused by the leakage of damaged macular vessels and choroid neovascularization are symptoms of many ophthalmic diseases. Optical coherence tomography (OCT), which provides clear 10-layer cross-sectional images of the retina, is widely used to screen various ophthalmic diseases. A large number of researchers have carried out relevant studies on deep learning technology to realize the semantic segmentation of lesion areas, such as effusion on OCT images, and achieved good results. However, in this field, problems of the low contrast of the lesion area and unevenness of lesion size limit the accuracy of the deep learning semantic segmentation model. In this paper, we propose a boundary multi-scale multi-task OCT segmentation network (BMM-Net) for these two challenges to segment the retinal edema area, subretinal fluid, and pigment epithelial detachment in OCT images. We propose a boundary extraction module, a multi-scale information perception module, and a classification module to capture accurate position and semantic information and collaboratively extract meaningful features. We train and verify on the AI Challenger competition dataset. The average Dice coefficient of the three lesion areas is 3.058% higher than the most commonly used model in the field of medical image segmentation and reaches 0.8222.
视网膜积液和囊肿是由受损的黄斑血管渗漏和脉络膜新生血管形成引起的,是许多眼病的症状。光学相干断层扫描(OCT)可提供清晰的视网膜10层横切面图像,被广泛用于各种眼科疾病的筛查。大量研究人员对深度学习技术进行了相关研究,实现病变区域的语义分割,如OCT图像上的积液,并取得了良好的效果。然而,在该领域中,病变区域对比度低、病变大小不均匀等问题限制了深度学习语义分割模型的准确性。针对这两种挑战,我们提出了一种边界多尺度多任务OCT分割网络(BMM-Net)来分割OCT图像中的视网膜水肿区、视网膜下液和色素上皮脱离。我们提出了边界提取模块、多尺度信息感知模块和分类模块,以捕获准确的位置和语义信息,协同提取有意义的特征。我们在AI挑战者比赛数据集上进行训练和验证。三个病灶区域的平均Dice系数比医学图像分割领域中最常用的模型高3.058%,达到0.8222。
{"title":"BMM-Net: automatic segmentation of edema in optical coherence tomography based on boundary detection and multi-scale network","authors":"Ruru Zhang, Jiawen He, Shenda Shi, E. Haihong, Zhonghong Ou, Meina Song","doi":"10.1145/3368555.3384447","DOIUrl":"https://doi.org/10.1145/3368555.3384447","url":null,"abstract":"Retinal effusions and cysts caused by the leakage of damaged macular vessels and choroid neovascularization are symptoms of many ophthalmic diseases. Optical coherence tomography (OCT), which provides clear 10-layer cross-sectional images of the retina, is widely used to screen various ophthalmic diseases. A large number of researchers have carried out relevant studies on deep learning technology to realize the semantic segmentation of lesion areas, such as effusion on OCT images, and achieved good results. However, in this field, problems of the low contrast of the lesion area and unevenness of lesion size limit the accuracy of the deep learning semantic segmentation model. In this paper, we propose a boundary multi-scale multi-task OCT segmentation network (BMM-Net) for these two challenges to segment the retinal edema area, subretinal fluid, and pigment epithelial detachment in OCT images. We propose a boundary extraction module, a multi-scale information perception module, and a classification module to capture accurate position and semantic information and collaboratively extract meaningful features. We train and verify on the AI Challenger competition dataset. The average Dice coefficient of the three lesion areas is 3.058% higher than the most commonly used model in the field of medical image segmentation and reaches 0.8222.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90166538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Using SNOMED to automate clinical concept mapping 使用SNOMED自动化临床概念映射
Pub Date : 2020-04-02 DOI: 10.1145/3368555.3384453
Shaun Gupta, Frederik Dieleman, P. Long, O. Doyle, N. Leavitt
The International Classification of Disease (ICD) is a widely used diagnostic ontology for the classification of health disorders and a valuable resource for healthcare analytics. However, ICD is an evolving ontology and subject to periodic revisions (e.g. ICD-9-CM to ICD-10-CM) resulting in the absence of complete cross-walks between versions. While clinical experts can create custom mappings across ICD versions, this process is both time-consuming and costly. We propose an automated solution that facilitates interoperability without sacrificing accuracy. Our solution leverages the SNOMED-CT ontology whereby medical concepts are organised in a directed acyclic graph. We use this to map ICD-9-CM to ICD-10-CM by associating codes to clinical concepts in the SNOMED graph using a nearest neighbors search in combination with natural language processing. To assess the impact of our method, the performance of a gradient boosted tree (XGBoost) developed to classify patients with Exocrine Pancreatic Insufficiency (EPI) disorder, was compared when using features constructed by our solution versus clinically-driven methods. This dataset comprised of 23, 204 EPI patients and 277, 324 non-EPI patients with data spanning from October 2011 to April 2017. Our algorithm generated clinical predictors with comparable stability across the ICD-9-CM to ICD-10-CM transition point when compared to ICD-9-CM/ICD-10-CM mappings generated by clinical experts. Preliminary modeling results showed highly similar performance for models based on the SNOMED mapping vs clinically defined mapping (71% precision at 20% recall for both models). Overall, the framework does not compromise on accuracy at the individual code level or at the model-level while obviating the need for time-consuming manual mapping.
国际疾病分类(ICD)是一个广泛使用的健康疾病分类诊断本体,也是医疗保健分析的宝贵资源。然而,ICD是一个不断发展的本体,并受到定期修订的影响(例如,ICD-9- cm到ICD-10- cm),导致版本之间缺乏完整的交叉行走。虽然临床专家可以跨ICD版本创建自定义映射,但这个过程既耗时又昂贵。我们提出一个自动化的解决方案,在不牺牲准确性的情况下促进互操作性。我们的解决方案利用了SNOMED-CT本体,将医学概念组织在一个有向无环图中。我们使用最近邻搜索结合自然语言处理,将代码与SNOMED图中的临床概念相关联,从而将ICD-9-CM映射到ICD-10-CM。为了评估我们的方法的影响,在使用我们的解决方案构建的特征与临床驱动的方法时,比较了用于对外分泌胰腺功能不全(EPI)疾病患者进行分类的梯度增强树(XGBoost)的性能。该数据集包括23,204名EPI患者和277,324名非EPI患者,数据时间跨度为2011年10月至2017年4月。与临床专家生成的ICD-9-CM/ICD-10-CM映射相比,我们的算法生成的临床预测因子在ICD-9-CM到ICD-10-CM的过渡点上具有相当的稳定性。初步的建模结果显示,基于SNOMED映射和临床定义映射的模型的性能非常相似(两种模型的准确率为71%,召回率为20%)。总的来说,框架在避免耗时的手工映射的同时,不会在单个代码级别或模型级别的准确性上做出妥协。
{"title":"Using SNOMED to automate clinical concept mapping","authors":"Shaun Gupta, Frederik Dieleman, P. Long, O. Doyle, N. Leavitt","doi":"10.1145/3368555.3384453","DOIUrl":"https://doi.org/10.1145/3368555.3384453","url":null,"abstract":"The International Classification of Disease (ICD) is a widely used diagnostic ontology for the classification of health disorders and a valuable resource for healthcare analytics. However, ICD is an evolving ontology and subject to periodic revisions (e.g. ICD-9-CM to ICD-10-CM) resulting in the absence of complete cross-walks between versions. While clinical experts can create custom mappings across ICD versions, this process is both time-consuming and costly. We propose an automated solution that facilitates interoperability without sacrificing accuracy. Our solution leverages the SNOMED-CT ontology whereby medical concepts are organised in a directed acyclic graph. We use this to map ICD-9-CM to ICD-10-CM by associating codes to clinical concepts in the SNOMED graph using a nearest neighbors search in combination with natural language processing. To assess the impact of our method, the performance of a gradient boosted tree (XGBoost) developed to classify patients with Exocrine Pancreatic Insufficiency (EPI) disorder, was compared when using features constructed by our solution versus clinically-driven methods. This dataset comprised of 23, 204 EPI patients and 277, 324 non-EPI patients with data spanning from October 2011 to April 2017. Our algorithm generated clinical predictors with comparable stability across the ICD-9-CM to ICD-10-CM transition point when compared to ICD-9-CM/ICD-10-CM mappings generated by clinical experts. Preliminary modeling results showed highly similar performance for models based on the SNOMED mapping vs clinically defined mapping (71% precision at 20% recall for both models). Overall, the framework does not compromise on accuracy at the individual code level or at the model-level while obviating the need for time-consuming manual mapping.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80046647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Proceedings of the ACM Conference on Health, Inference, and Learning
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1