首页 > 最新文献

Artificial intelligence in the life sciences最新文献

英文 中文
Yoked learning in molecular data science 分子数据科学中的交配学习
Pub Date : 2023-12-02 DOI: 10.1016/j.ailsci.2023.100089
Zhixiong Li, Yan Xiang, Yujing Wen, Daniel Reker

Active machine learning is an established and increasingly popular experimental design technique where the machine learning model can request additional data to improve the model's predictive performance. It is generally assumed that this data is optimal for the machine learning model since it relies on the model's predictions or model architecture and therefore cannot be transferred to other models. Inspired by research in pedagogy, we here introduce the concept of yoked machine learning where a second machine learning model learns from the data selected by another model. We found that in 48% of the benchmarked combinations, yoked learning performed similar or better than active learning. We analyze distinct cases in which yoked learning can improve active learning performance. In particular, we prototype yoked deep learning (YoDeL) where a classic machine learning model provides data to a deep neural network, thereby mitigating challenges of active deep learning such as slow refitting time per learning iteration and poor performance on small datasets. In summary, we expect the new concept of yoked (deep) learning to provide a competitive option to boost the performance of active learning and benefit from distinct capabilities of multiple machine learning models during data acquisition, training, and deployment.

主动式机器学习是一种成熟且日益流行的实验设计技术,机器学习模型可以请求额外的数据来提高模型的预测性能。一般认为,这些数据是机器学习模型的最佳数据,因为这些数据依赖于模型的预测或模型架构,因此不能转移到其他模型中。受教学法研究的启发,我们在此引入了枷锁式机器学习的概念,即第二个机器学习模型从另一个模型选择的数据中学习。我们发现,在 48% 的基准组合中,连带学习的表现与主动学习相似或更好。我们分析了联合学习可以提高主动学习性能的不同情况。特别是,我们提出了枷锁式深度学习(YoDeL)的原型,即由一个经典机器学习模型为深度神经网络提供数据,从而缓解主动式深度学习所面临的挑战,如每次学习迭代的重拟合时间较慢以及在小数据集上的性能较差。总之,我们希望轭状(深度)学习这一新概念能提供一种有竞争力的选择,以提高主动学习的性能,并在数据采集、训练和部署过程中受益于多种机器学习模型的独特能力。
{"title":"Yoked learning in molecular data science","authors":"Zhixiong Li,&nbsp;Yan Xiang,&nbsp;Yujing Wen,&nbsp;Daniel Reker","doi":"10.1016/j.ailsci.2023.100089","DOIUrl":"https://doi.org/10.1016/j.ailsci.2023.100089","url":null,"abstract":"<div><p>Active machine learning is an established and increasingly popular experimental design technique where the machine learning model can request additional data to improve the model's predictive performance. It is generally assumed that this data is optimal for the machine learning model since it relies on the model's predictions or model architecture and therefore cannot be transferred to other models. Inspired by research in pedagogy, we here introduce the concept of yoked machine learning where a second machine learning model learns from the data selected by another model. We found that in 48% of the benchmarked combinations, yoked learning performed similar or better than active learning. We analyze distinct cases in which yoked learning can improve active learning performance. In particular, we prototype yoked deep learning (YoDeL) where a classic machine learning model provides data to a deep neural network, thereby mitigating challenges of active deep learning such as slow refitting time per learning iteration and poor performance on small datasets. In summary, we expect the new concept of yoked (deep) learning to provide a competitive option to boost the performance of active learning and benefit from distinct capabilities of multiple machine learning models during data acquisition, training, and deployment.</p></div>","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"5 ","pages":"Article 100089"},"PeriodicalIF":0.0,"publicationDate":"2023-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2667318523000338/pdfft?md5=798e4cffb7539da96cce07297e51e3de&pid=1-s2.0-S2667318523000338-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138570365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A supervised machine learning workflow for the reduction of highly dimensional biological data 用于减少高维生物数据的有监督机器学习工作流程
Pub Date : 2023-11-25 DOI: 10.1016/j.ailsci.2023.100090
Linnea K. Andersen , Benjamin J. Reading

Recent technological advancements have revolutionized research capabilities across the biological sciences by enabling the collection of large data that provides a broader picture of systems from the cellular to ecosystem level at a more refined resolution. The rapid rate of generating these data has exacerbated bottlenecks in study design and data analysis approaches, especially as conventional methods that incorporate traditional statistical tests and assumptions are not suitable or sufficient for highly dimensional data (i.e., more than 1,000 variables). The application of machine learning techniques in large data analysis is one promising solution that is increasingly popular. However, limitations in expertise such that the results from machine learning models can be interpreted to gain meaningful biological insight pose a great challenge. To address this challenge, a user-friendly machine learning workflow that can be applied to a wide variety of data types to reduce these large data to those variables (attributes) most determinant of experimental and/or observed conditions is provided, as well as a general overview of data analysis and machine learning approaches and considerations thereof. The workflow presented here has been beta-tested with great success and is recommended to be incorporated into analysis pipelines of large data as a standardized approach to reduce data dimensionality. Moreover, the workflow is flexible, and the underlying concepts and steps can be modified to best suit user needs, objectives, and study parameters.

最近的技术进步彻底改变了整个生物科学领域的研究能力,使我们能够收集大量数据,以更精细的分辨率提供从细胞到生态系统层面的更广阔的系统图景。这些数据的快速生成加剧了研究设计和数据分析方法的瓶颈,尤其是包含传统统计检验和假设的传统方法不适合或不足以处理高维数据(即超过 1,000 个变量)。在大数据分析中应用机器学习技术是一种很有前景的解决方案,而且越来越受欢迎。然而,由于专业知识的限制,如何解释机器学习模型的结果以获得有意义的生物学见解成为一个巨大的挑战。为了应对这一挑战,本文提供了一个用户友好型机器学习工作流程,该流程可应用于多种数据类型,将这些海量数据还原为对实验和/或观测条件最具决定性的变量(属性),同时还概述了数据分析和机器学习方法及其注意事项。本文介绍的工作流程已经过测试,取得了巨大成功,建议将其纳入大数据分析管道,作为降低数据维度的标准化方法。此外,该工作流程非常灵活,可根据用户需求、目标和研究参数对基本概念和步骤进行修改。
{"title":"A supervised machine learning workflow for the reduction of highly dimensional biological data","authors":"Linnea K. Andersen ,&nbsp;Benjamin J. Reading","doi":"10.1016/j.ailsci.2023.100090","DOIUrl":"https://doi.org/10.1016/j.ailsci.2023.100090","url":null,"abstract":"<div><p>Recent technological advancements have revolutionized research capabilities across the biological sciences by enabling the collection of large data that provides a broader picture of systems from the cellular to ecosystem level at a more refined resolution. The rapid rate of generating these data has exacerbated bottlenecks in study design and data analysis approaches, especially as conventional methods that incorporate traditional statistical tests and assumptions are not suitable or sufficient for highly dimensional data (i.e., more than 1,000 variables). The application of machine learning techniques in large data analysis is one promising solution that is increasingly popular. However, limitations in expertise such that the results from machine learning models can be interpreted to gain meaningful biological insight pose a great challenge. To address this challenge, a user-friendly machine learning workflow that can be applied to a wide variety of data types to reduce these large data to those variables (attributes) most determinant of experimental and/or observed conditions is provided, as well as a general overview of data analysis and machine learning approaches and considerations thereof. The workflow presented here has been beta-tested with great success and is recommended to be incorporated into analysis pipelines of large data as a standardized approach to reduce data dimensionality. Moreover, the workflow is flexible, and the underlying concepts and steps can be modified to best suit user needs, objectives, and study parameters.</p></div>","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"5 ","pages":"Article 100090"},"PeriodicalIF":0.0,"publicationDate":"2023-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S266731852300034X/pdfft?md5=c41b31c74fb0a867fbb87db01c8f6190&pid=1-s2.0-S266731852300034X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138739061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
First-generation themed article collections 第一代主题文集
Pub Date : 2023-11-15 DOI: 10.1016/j.ailsci.2023.100088
Jürgen Bajorath, Steve Gardner, Francesca Grisoni, Carolina Horta Andrade, Johannes Kirchmair, Melissa Landon, José L. Medina-Franco, Filip Miljković, Floriane Montantari, Raquel Rodríguez-Pérez
{"title":"First-generation themed article collections","authors":"Jürgen Bajorath,&nbsp;Steve Gardner,&nbsp;Francesca Grisoni,&nbsp;Carolina Horta Andrade,&nbsp;Johannes Kirchmair,&nbsp;Melissa Landon,&nbsp;José L. Medina-Franco,&nbsp;Filip Miljković,&nbsp;Floriane Montantari,&nbsp;Raquel Rodríguez-Pérez","doi":"10.1016/j.ailsci.2023.100088","DOIUrl":"https://doi.org/10.1016/j.ailsci.2023.100088","url":null,"abstract":"","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"4 ","pages":"Article 100088"},"PeriodicalIF":0.0,"publicationDate":"2023-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2667318523000326/pdfft?md5=cf39af5e517634e52f28e066ee1fd1da&pid=1-s2.0-S2667318523000326-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136697050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Experimental Uncertainty in Training Data for Protein-Ligand Binding Affinity Prediction Models 蛋白质配体结合亲和力预测模型训练数据的实验不确定性
Pub Date : 2023-10-04 DOI: 10.1016/j.ailsci.2023.100087
Carlos A. Hernández-Garrido , Norberto Sánchez-Cruz

The accuracy of machine learning models for protein-ligand binding affinity prediction depends on the quality of the experimental data they are trained on. Most of these models are trained and tested on different subsets of the PDBbind database, which is the main source of protein-ligand complexes with annotated binding affinity in the public domain. However, estimating its experimental uncertainty is not straightforward because just a few protein-ligand complexes have more than one measurement associated. In this work, we analyze bioactivity data from ChEMBL to estimate the experimental uncertainty associated with the three binding affinity measures included in the PDBbind (Ki, Kd, and IC50), as well as the effect of combining them. The experimental uncertainty of combining these three affinity measures was characterized by a mean absolute error of 0.78 logarithmic units, a root mean square error of 1.04 and a Pearson correlation coefficient of 0.76. These estimations were contrasted with the performances obtained by state-of-the-art machine learning models for binding affinity prediction, showing that these models tend to be overoptimistic when evaluated on the core set from PDBbind.

用于蛋白质-配体结合亲和力预测的机器学习模型的准确性取决于它们所训练的实验数据的质量。这些模型中的大多数都是在PDBbind数据库的不同子集上训练和测试的,PDBBinding数据库是公共领域中具有注释结合亲和力的蛋白质-配体复合物的主要来源。然而,估计其实验不确定性并不简单,因为只有少数蛋白质-配体复合物具有多个相关测量。在这项工作中,我们分析了来自ChEMBL的生物活性数据,以估计与PDBbind中包括的三种结合亲和力测量(Ki、Kd和IC50)相关的实验不确定性,以及将它们结合的效果。组合这三种亲和性测量的实验不确定度的特征是平均绝对误差为0.78对数单位,均方根误差为1.04,Pearson相关系数为0.76。这些估计与用于结合亲和力预测的最先进的机器学习模型获得的性能进行了对比,表明当在PDBbind的核心集上进行评估时,这些模型往往过于乐观。
{"title":"Experimental Uncertainty in Training Data for Protein-Ligand Binding Affinity Prediction Models","authors":"Carlos A. Hernández-Garrido ,&nbsp;Norberto Sánchez-Cruz","doi":"10.1016/j.ailsci.2023.100087","DOIUrl":"https://doi.org/10.1016/j.ailsci.2023.100087","url":null,"abstract":"<div><p>The accuracy of machine learning models for protein-ligand binding affinity prediction depends on the quality of the experimental data they are trained on. Most of these models are trained and tested on different subsets of the PDBbind database, which is the main source of protein-ligand complexes with annotated binding affinity in the public domain. However, estimating its experimental uncertainty is not straightforward because just a few protein-ligand complexes have more than one measurement associated. In this work, we analyze bioactivity data from ChEMBL to estimate the experimental uncertainty associated with the three binding affinity measures included in the PDBbind (K<sub>i</sub>, K<sub>d</sub>, and IC<sub>50</sub>), as well as the effect of combining them. The experimental uncertainty of combining these three affinity measures was characterized by a mean absolute error of 0.78 logarithmic units, a root mean square error of 1.04 and a Pearson correlation coefficient of 0.76. These estimations were contrasted with the performances obtained by state-of-the-art machine learning models for binding affinity prediction, showing that these models tend to be overoptimistic when evaluated on the core set from PDBbind.</p></div>","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"4 ","pages":"Article 100087"},"PeriodicalIF":0.0,"publicationDate":"2023-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49711349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring new horizons: Empowering computer-assisted drug design with few-shot learning 探索新视野:通过少量的注射学习实现计算机辅助药物设计
Pub Date : 2023-09-09 DOI: 10.1016/j.ailsci.2023.100086
Sabrina Silva-Mendonça , Arthur Ricardo de Sousa Vitória , Telma Woerle de Lima , Arlindo Rodrigues Galvão-Filho , Carolina Horta Andrade

Computational approaches have revolutionized the field of drug discovery, collectively known as Computer-Assisted Drug Design (CADD). Advancements in computing power, data generation, digitalization, and artificial intelligence (AI) techniques have played a crucial role in the rise of CADD. These approaches offer numerous benefits, enabling the analysis and interpretation of vast amounts of data from diverse sources, such as genomics, structural information, and clinical trials data. By integrating and analyzing these multiple data sources, researchers can efficiently identify potential drug targets and develop new drug candidates. Among the AI techniques, machine learning (ML) and deep learning (DL) have shown tremendous promise in drug discovery. ML and DL models can effectively utilize experimental data to accurately predict the efficacy and safety of drug candidates. However, despite these advancements, certain areas in drug discovery face data scarcity, particularly in neglected, rare, and emerging viral diseases. Few-shot learning (FSL) is an emerging approach that addresses the challenge of limited data in drug discovery. FSL enables ML models to learn from a small number of examples of a new task, achieving commendable performance by leveraging knowledge learned from related datasets or prior information. It often involves meta-learning, which trains a model to learn how to learn from few data. This ability to quickly adapt to new tasks with low data circumvents the need for extensive training on large datasets. By enabling efficient learning from a small amount of data, few-shot learning has the potential to accelerate the drug discovery process and enhance the success rate of drug development. In this review, we introduce the concept of few-shot learning and its application in drug discovery. Furthermore, we demonstrate the valuable application of few-shot learning in the identification of new drug targets, accurate prediction of drug efficacy, and the design of novel compounds possessing desired biological properties. This comprehensive review draws upon numerous papers from the literature to provide extensive insights into the effectiveness and potential of few-shot learning in these critical areas of drug discovery and development.

计算方法彻底改变了药物发现领域,统称为计算机辅助药物设计(CADD)。计算能力、数据生成、数字化和人工智能(AI)技术的进步在CADD的兴起中发挥了至关重要的作用。这些方法提供了许多好处,能够分析和解释来自不同来源的大量数据,如基因组学、结构信息和临床试验数据。通过整合和分析这些多个数据源,研究人员可以有效地识别潜在的药物靶点并开发新的候选药物。在人工智能技术中,机器学习(ML)和深度学习(DL)在药物发现方面显示出巨大的前景。ML和DL模型可以有效地利用实验数据来准确预测候选药物的疗效和安全性。然而,尽管取得了这些进展,药物发现的某些领域仍面临数据短缺,尤其是在被忽视、罕见和新出现的病毒性疾病方面。少量注射学习(FSL)是一种新兴的方法,可以解决药物发现中数据有限的挑战。FSL使ML模型能够从新任务的少量示例中学习,通过利用从相关数据集或先前信息中学习的知识实现了值得称赞的性能。它通常涉及元学习,它训练模型学习如何从少量数据中学习。这种快速适应低数据新任务的能力避免了在大型数据集上进行广泛训练的需要。通过从少量数据中实现高效学习,少镜头学习有可能加速药物发现过程,提高药物开发的成功率。在这篇综述中,我们介绍了少镜头学习的概念及其在药物发现中的应用。此外,我们展示了少镜头学习在识别新药靶点、准确预测药效以及设计具有所需生物特性的新型化合物方面的宝贵应用。这篇全面的综述借鉴了文献中的大量论文,对少针学习在药物发现和开发的这些关键领域的有效性和潜力提供了广泛的见解。
{"title":"Exploring new horizons: Empowering computer-assisted drug design with few-shot learning","authors":"Sabrina Silva-Mendonça ,&nbsp;Arthur Ricardo de Sousa Vitória ,&nbsp;Telma Woerle de Lima ,&nbsp;Arlindo Rodrigues Galvão-Filho ,&nbsp;Carolina Horta Andrade","doi":"10.1016/j.ailsci.2023.100086","DOIUrl":"https://doi.org/10.1016/j.ailsci.2023.100086","url":null,"abstract":"<div><p>Computational approaches have revolutionized the field of drug discovery, collectively known as Computer-Assisted Drug Design (CADD). Advancements in computing power, data generation, digitalization, and artificial intelligence (AI) techniques have played a crucial role in the rise of CADD. These approaches offer numerous benefits, enabling the analysis and interpretation of vast amounts of data from diverse sources, such as genomics, structural information, and clinical trials data. By integrating and analyzing these multiple data sources, researchers can efficiently identify potential drug targets and develop new drug candidates. Among the AI techniques, machine learning (ML) and deep learning (DL) have shown tremendous promise in drug discovery. ML and DL models can effectively utilize experimental data to accurately predict the efficacy and safety of drug candidates. However, despite these advancements, certain areas in drug discovery face data scarcity, particularly in neglected, rare, and emerging viral diseases. Few-shot learning (FSL) is an emerging approach that addresses the challenge of limited data in drug discovery. FSL enables ML models to learn from a small number of examples of a new task, achieving commendable performance by leveraging knowledge learned from related datasets or prior information. It often involves meta-learning, which trains a model to learn how to learn from few data. This ability to quickly adapt to new tasks with low data circumvents the need for extensive training on large datasets. By enabling efficient learning from a small amount of data, few-shot learning has the potential to accelerate the drug discovery process and enhance the success rate of drug development. In this review, we introduce the concept of few-shot learning and its application in drug discovery. Furthermore, we demonstrate the valuable application of few-shot learning in the identification of new drug targets, accurate prediction of drug efficacy, and the design of novel compounds possessing desired biological properties. This comprehensive review draws upon numerous papers from the literature to provide extensive insights into the effectiveness and potential of few-shot learning in these critical areas of drug discovery and development.</p></div>","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"4 ","pages":"Article 100086"},"PeriodicalIF":0.0,"publicationDate":"2023-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49711348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Data and code availability requirements in open science and consequences for different research environments 开放科学中的数据和代码可用性要求及其对不同研究环境的影响
Pub Date : 2023-08-23 DOI: 10.1016/j.ailsci.2023.100085
Jürgen Bajorath
{"title":"Data and code availability requirements in open science and consequences for different research environments","authors":"Jürgen Bajorath","doi":"10.1016/j.ailsci.2023.100085","DOIUrl":"10.1016/j.ailsci.2023.100085","url":null,"abstract":"","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"4 ","pages":"Article 100085"},"PeriodicalIF":0.0,"publicationDate":"2023-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47255704","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analysis of Swin-UNet vision transformer for Inferior Vena Cava filter segmentation from CT scans Swin-UNet视觉变换器用于下腔静脉CT滤波分割的分析
Pub Date : 2023-08-18 DOI: 10.1016/j.ailsci.2023.100084
Rahul Gomes , Tyler Pham , Nichol He , Connor Kamrowski , Joseph Wildenberg

Purpose

The purpose of this study is to develop an accurate deep learning model capable of Inferior Vena Cava (IVC) filter segmentation from CT scans. The study does a comparative assessment of the impact of Residual Networks (ResNets) complemented with reduced convolutional layer depth and also analyzes the impact of using vision transformer architectures without performance degradation.

Materials and Methods

This experimental retrospective study on 84 CT scans consisting of 54618 slices involves design, implementation, and evaluation of segmentation algorithm which can be used to generate a clinical report for the presence of IVC filters on abdominal CT scans performed for any reason. Several variants of patch-based 3D-Convolutional Neural Network (CNN) and the Swin UNet Transformer (Swin-UNETR) are used to retrieve the signature of IVC filters. The Dice Score is used as a metric to compare the performance of the segmentation models.

Results

Model trained on UNet variant using four ResNet layers showed a higher segmentation performance achieving median Dice = 0.92 [Interquartile range(IQR): 0.85, 0.93] compared to the plain UNet model with four layers having median Dice = 0.89 [IQR: 0.83, 0.92]. Segmentation results from ResNet with two layers achieved a median Dice = 0.93 [IQR: 0.87, 0.94] which was higher than the plain UNet model with two layers at median Dice = 0.87 [IQR: 0.77, 0.90]. Models trained using SWIN-based transformers performed significantly better in both training and validation datasets compared to the four CNN variants. The validation median Dice was highest in 4 layer Swin UNETR at 0.88 followed by 2 layer Swin UNETR at 0.85.

Conclusion

Utilization of vision based transformer Swin-UNETR results in segmentation output with both low bias and variance thereby solving a real-world problem within healthcare for advanced Artificial Intelligence (AI) image processing and recognition. The Swin UNETR will reduce the time spent manually tracking IVC filters by centralizing within the electronic health record. Link to GitHub repository.

目的建立一种精确的深度学习模型,用于下腔静脉(IVC) CT图像的滤波分割。该研究对残差网络(ResNets)与减少卷积层深度相结合的影响进行了比较评估,并分析了在不降低性能的情况下使用视觉转换器架构的影响。材料和方法本实验回顾性研究了84个CT扫描,包括54618个切片,涉及分割算法的设计、实现和评估,该算法可用于生成临床报告,用于任何原因进行的腹部CT扫描中存在IVC过滤器。基于补丁的三维卷积神经网络(CNN)和Swin UNet变压器(swan - unetr)的几种变体被用于检索IVC滤波器的特征。Dice Score被用作比较分割模型性能的指标。结果使用4个ResNet层训练的UNet变体模型与使用4个ResNet层训练的UNet模型相比,具有更高的分割性能,达到中位数Dice = 0.92[四分位间距(IQR): 0.85, 0.93],而普通UNet模型的中位数Dice = 0.89 [IQR: 0.83, 0.92]。ResNet两层分割结果的中位数Dice = 0.93 [IQR: 0.87, 0.94],高于普通UNet两层模型的中位数Dice = 0.87 [IQR: 0.77, 0.90]。与四种CNN变体相比,使用基于swn的变压器训练的模型在训练和验证数据集中的表现都要好得多。4层Swin UNETR的验证中位数骰子最高,为0.88,其次是2层Swin UNETR,为0.85。结论使用基于视觉的swun - unetr变压器可以获得低偏差和方差的分割输出,从而解决了先进人工智能(AI)图像处理和识别在医疗保健中的现实问题。Swin UNETR将通过集中在电子健康记录内减少人工跟踪IVC过滤器所花费的时间。链接到GitHub仓库。
{"title":"Analysis of Swin-UNet vision transformer for Inferior Vena Cava filter segmentation from CT scans","authors":"Rahul Gomes ,&nbsp;Tyler Pham ,&nbsp;Nichol He ,&nbsp;Connor Kamrowski ,&nbsp;Joseph Wildenberg","doi":"10.1016/j.ailsci.2023.100084","DOIUrl":"10.1016/j.ailsci.2023.100084","url":null,"abstract":"<div><h3>Purpose</h3><p>The purpose of this study is to develop an accurate deep learning model capable of Inferior Vena Cava (IVC) filter segmentation from CT scans. The study does a comparative assessment of the impact of Residual Networks (ResNets) complemented with reduced convolutional layer depth and also analyzes the impact of using vision transformer architectures without performance degradation.</p></div><div><h3>Materials and Methods</h3><p>This experimental retrospective study on 84 CT scans consisting of 54618 slices involves design, implementation, and evaluation of segmentation algorithm which can be used to generate a clinical report for the presence of IVC filters on abdominal CT scans performed for any reason. Several variants of patch-based 3D-Convolutional Neural Network (CNN) and the Swin UNet Transformer (Swin-UNETR) are used to retrieve the signature of IVC filters. The Dice Score is used as a metric to compare the performance of the segmentation models.</p></div><div><h3>Results</h3><p>Model trained on UNet variant using four ResNet layers showed a higher segmentation performance achieving median Dice = 0.92 [Interquartile range(IQR): 0.85, 0.93] compared to the plain UNet model with four layers having median Dice = 0.89 [IQR: 0.83, 0.92]. Segmentation results from ResNet with two layers achieved a median Dice = 0.93 [IQR: 0.87, 0.94] which was higher than the plain UNet model with two layers at median Dice = 0.87 [IQR: 0.77, 0.90]. Models trained using SWIN-based transformers performed significantly better in both training and validation datasets compared to the four CNN variants. The validation median Dice was highest in 4 layer Swin UNETR at 0.88 followed by 2 layer Swin UNETR at 0.85.</p></div><div><h3>Conclusion</h3><p>Utilization of vision based transformer Swin-UNETR results in segmentation output with both low bias and variance thereby solving a real-world problem within healthcare for advanced Artificial Intelligence (AI) image processing and recognition. The Swin UNETR will reduce the time spent manually tracking IVC filters by centralizing within the electronic health record. Link to <span>GitHub</span><svg><path></path></svg> repository.</p></div>","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"4 ","pages":"Article 100084"},"PeriodicalIF":0.0,"publicationDate":"2023-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46348564","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep neural network architectures for cardiac image segmentation 用于心脏图像分割的深度神经网络结构
Pub Date : 2023-08-09 DOI: 10.1016/j.ailsci.2023.100083
Jasmine El-Taraboulsi , Claudia P. Cabrera , Caroline Roney , Nay Aung

Imaging plays a fundamental role in the effective diagnosis, staging, management, and monitoring of various cardiac pathologies. Successful radiological analysis relies on accurate image segmentation, a technically arduous process, prone to human-error. To overcome the laborious and time-consuming nature of cardiac image analysis, deep learning approaches have been developed, enabling the accurate, time-efficient, and highly personalised diagnosis, staging and management of cardiac pathologies.

Here, we present a review of over 60 papers, proposing deep learning models for cardiac image segmentation. We summarise the theoretical basis of Convolutional Neural Networks, Fully Convolutional Neural Networks, U-Net, V-Net, No-New-U-Net (nnU-Net), Transformer Networks, DeepLab, Generative Adversarial Networks, Auto Encoders and Recurrent Neural Networks. In addition, we identify pertinent performance-enhancing measures including adaptive convolutional kernels, atrous convolutions, attention gates, and deep supervision modules.

Top-performing models in ventricular, myocardial, atrial and aortic segmentation are explored, highlighting U-Net and nnU-Net-based model architectures achieving state-of-the art segmentation accuracies. Additionally, key gaps in the current research and technology are identified, and areas of future research are suggested, aiming to guide the innovation and clinical adoption of automated cardiac segmentation methods.

影像在各种心脏疾病的有效诊断、分期、管理和监测中起着重要作用。成功的放射分析依赖于准确的图像分割,这是一个技术上艰巨的过程,容易出现人为错误。为了克服心脏图像分析的费力和耗时的性质,深度学习方法已经被开发出来,能够准确,高效,高度个性化的心脏病理诊断,分期和管理。在这里,我们回顾了60多篇论文,提出了用于心脏图像分割的深度学习模型。我们总结了卷积神经网络、全卷积神经网络、U-Net、V-Net、No-New-U-Net (nnU-Net)、变压器网络、DeepLab、生成对抗网络、自动编码器和循环神经网络的理论基础。此外,我们还确定了相关的性能增强措施,包括自适应卷积核、亚属性卷积、注意门和深度监督模块。探索了心室、心肌、心房和主动脉分割中表现最好的模型,突出了基于U-Net和nnu - net的模型架构,实现了最先进的分割精度。此外,指出了当前研究和技术的关键差距,并提出了未来的研究领域,旨在指导自动化心脏分割方法的创新和临床应用。
{"title":"Deep neural network architectures for cardiac image segmentation","authors":"Jasmine El-Taraboulsi ,&nbsp;Claudia P. Cabrera ,&nbsp;Caroline Roney ,&nbsp;Nay Aung","doi":"10.1016/j.ailsci.2023.100083","DOIUrl":"10.1016/j.ailsci.2023.100083","url":null,"abstract":"<div><p>Imaging plays a fundamental role in the effective diagnosis, staging, management, and monitoring of various cardiac pathologies. Successful radiological analysis relies on accurate image segmentation, a technically arduous process, prone to human-error. To overcome the laborious and time-consuming nature of cardiac image analysis, deep learning approaches have been developed, enabling the accurate, time-efficient, and highly personalised diagnosis, staging and management of cardiac pathologies.</p><p>Here, we present a review of over 60 papers, proposing deep learning models for cardiac image segmentation. We summarise the theoretical basis of Convolutional Neural Networks, Fully Convolutional Neural Networks, U-Net, V-Net, No-New-U-Net (nnU-Net), Transformer Networks, DeepLab, Generative Adversarial Networks, Auto Encoders and Recurrent Neural Networks. In addition, we identify pertinent performance-enhancing measures including adaptive convolutional kernels, atrous convolutions, attention gates, and deep supervision modules.</p><p>Top-performing models in ventricular, myocardial, atrial and aortic segmentation are explored, highlighting U-Net and nnU-Net-based model architectures achieving state-of-the art segmentation accuracies. Additionally, key gaps in the current research and technology are identified, and areas of future research are suggested, aiming to guide the innovation and clinical adoption of automated cardiac segmentation methods.</p></div>","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"4 ","pages":"Article 100083"},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45888732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modeling and survival exploration of breast carcinoma: A statistical, maximum likelihood estimation, and artificial neural network perspective 乳腺癌的建模和生存探索:统计学、最大似然估计和人工神经网络的视角
Pub Date : 2023-07-17 DOI: 10.1016/j.ailsci.2023.100082
Anum Shafiq , Andaç Batur Çolak , Tabassum Naz Sindhu , Showkat Ahmad Lone , Tahani A. Abushal

The core objective of this research is to describe the behavior of the distribution using the MLE method to estimate its parameters, as well as to determine the optimal Artificial Neural Network method by comparing it to the maximum likelihood estimation method and applying it to real data for breast cancer patients to determine survival, risk, and other survival study functions of the log-logistic distribution. The parameters were defined in the input layer of the artificial neural network developed for the purpose of survival analysis and reliability function, hazard rate function, probability density function, reserved hazard rate function, Mills ratio, Odd function and CHR values were obtained in the output layer. The findings show that risk function increases with the increase in the time of infection and then decreases for a group of breast cancer patients under study, which corresponds to the theoretical properties of this according to the practical conclusions. The examination of survival analysis reveals that practical conclusions correspond to the theoretical properties of log-logistic distribution. Artificial neural networks have proven to be one of the ideal tools that can be used to predict various vital parameters, especially survival of cancer patients, with their high predictive capabilities.

本研究的核心目标是利用最大似然估计方法来描述分布的行为,并将其与最大似然估计方法进行比较,并将其应用于乳腺癌患者的实际数据,确定log-logistic分布的生存、风险等生存研究函数,从而确定最优的人工神经网络方法。在为生存分析而开发的人工神经网络的输入层定义参数,并在输出层获得可靠性函数、风险率函数、概率密度函数、保留风险率函数、Mills比、Odd函数和CHR值。研究结果表明,在所研究的一组乳腺癌患者中,风险函数随着感染时间的增加而增加,然后降低,这与实际结论的理论性质相对应。对生存分析的检验表明,实际结论符合逻辑-logistic分布的理论性质。人工神经网络已被证明是预测各种重要参数,特别是癌症患者生存的理想工具之一,具有很高的预测能力。
{"title":"Modeling and survival exploration of breast carcinoma: A statistical, maximum likelihood estimation, and artificial neural network perspective","authors":"Anum Shafiq ,&nbsp;Andaç Batur Çolak ,&nbsp;Tabassum Naz Sindhu ,&nbsp;Showkat Ahmad Lone ,&nbsp;Tahani A. Abushal","doi":"10.1016/j.ailsci.2023.100082","DOIUrl":"10.1016/j.ailsci.2023.100082","url":null,"abstract":"<div><p>The core objective of this research is to describe the behavior of the distribution using the MLE method to estimate its parameters, as well as to determine the optimal Artificial Neural Network method by comparing it to the maximum likelihood estimation method and applying it to real data for breast cancer patients to determine survival, risk, and other survival study functions of the log-logistic distribution. The parameters were defined in the input layer of the artificial neural network developed for the purpose of survival analysis and reliability function, hazard rate function, probability density function, reserved hazard rate function, Mills ratio, Odd function and CHR values were obtained in the output layer. The findings show that risk function increases with the increase in the time of infection and then decreases for a group of breast cancer patients under study, which corresponds to the theoretical properties of this according to the practical conclusions. The examination of survival analysis reveals that practical conclusions correspond to the theoretical properties of log-logistic distribution. Artificial neural networks have proven to be one of the ideal tools that can be used to predict various vital parameters, especially survival of cancer patients, with their high predictive capabilities.</p></div>","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"4 ","pages":"Article 100082"},"PeriodicalIF":0.0,"publicationDate":"2023-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41536157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
piCRISPR: Physically informed deep learning models for CRISPR/Cas9 off-target cleavage prediction piCRISPR:用于CRISPR/Cas9脱靶切割预测的物理信息深度学习模型
Pub Date : 2023-05-15 DOI: 10.1016/j.ailsci.2023.100075
Florian Störtz, Jeffrey K. Mak, Peter Minary

CRISPR/Cas programmable nuclease systems have become ubiquitous in the field of gene editing. With progressing development, applications in in vivo therapeutic gene editing are increasingly within reach, yet limited by possible adverse side effects from unwanted edits. Recent years have thus seen continuous development of off-target prediction algorithms trained on in vitro cleavage assay data gained from immortalised cell lines. It has been shown that in contrast to experimental epigenetic features, computed physically informed features are so far underutilised despite bearing considerably larger correlation with cleavage activity. Here, we implement state-of-the-art deep learning algorithms and feature encodings for off-target prediction with emphasis on physically informed features that capture the biological environment of the cleavage site, hence terming our approach piCRISPR. Features were gained from the large, diverse crisprSQL off-target cleavage dataset. We find that our best-performing models highlight the importance of sequence context and chromatin accessibility for cleavage prediction and compare favourably with literature standard prediction performance. We further show that our novel, environmentally sensitive features are crucial to accurate prediction on sequence-identical locus pairs, making them highly relevant for clinical guide design. The source code and trained models can be found ready to use at github.com/florianst/picrispr.

CRISPR/Cas可编程核酸酶系统在基因编辑领域已经变得无处不在。随着开发的进展,体内治疗性基因编辑的应用越来越触手可及,但受到不必要编辑可能产生的副作用的限制。因此,近年来,在从永生化细胞系获得的体外切割测定数据上训练的脱靶预测算法不断发展。研究表明,与实验表观遗传学特征相比,尽管计算的物理信息特征与切割活性具有相当大的相关性,但迄今为止尚未得到充分利用。在这里,我们实现了最先进的深度学习算法和特征编码,用于脱靶预测,重点是捕捉切割位点生物环境的物理信息特征,从而确定了我们的方法piCRISPR。特征是从大型、多样化的crisprSQL脱靶切割数据集中获得的。我们发现,我们表现最好的模型强调了序列上下文和染色质可及性对切割预测的重要性,并与文献标准预测性能相比较。我们进一步表明,我们新颖的环境敏感特征对于准确预测序列相同的基因座对至关重要,使其与临床指南设计高度相关。源代码和经过训练的模型可以在github.com/florians/picrispr上找到。
{"title":"piCRISPR: Physically informed deep learning models for CRISPR/Cas9 off-target cleavage prediction","authors":"Florian Störtz,&nbsp;Jeffrey K. Mak,&nbsp;Peter Minary","doi":"10.1016/j.ailsci.2023.100075","DOIUrl":"https://doi.org/10.1016/j.ailsci.2023.100075","url":null,"abstract":"<div><p>CRISPR/Cas programmable nuclease systems have become ubiquitous in the field of gene editing. With progressing development, applications in <em>in vivo</em> therapeutic gene editing are increasingly within reach, yet limited by possible adverse side effects from unwanted edits. Recent years have thus seen continuous development of off-target prediction algorithms trained on <em>in vitro</em> cleavage assay data gained from immortalised cell lines. It has been shown that in contrast to experimental epigenetic features, computed physically informed features are so far underutilised despite bearing considerably larger correlation with cleavage activity. Here, we implement state-of-the-art deep learning algorithms and feature encodings for off-target prediction with emphasis on <em>physically informed</em> features that capture the biological environment of the cleavage site, hence terming our approach piCRISPR. Features were gained from the large, diverse crisprSQL off-target cleavage dataset. We find that our best-performing models highlight the importance of sequence context and chromatin accessibility for cleavage prediction and compare favourably with literature standard prediction performance. We further show that our novel, environmentally sensitive features are crucial to accurate prediction on sequence-identical locus pairs, making them highly relevant for clinical guide design. The source code and trained models can be found ready to use at <span>github.com/florianst/picrispr</span><svg><path></path></svg>.</p></div>","PeriodicalId":72304,"journal":{"name":"Artificial intelligence in the life sciences","volume":"3 ","pages":"Article 100075"},"PeriodicalIF":0.0,"publicationDate":"2023-05-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49774977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial intelligence in the life sciences
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1