首页 > 最新文献

ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine最新文献

英文 中文
Fast and memory-efficient scRNA-seq k-means clustering with various distances. 快速和高效的scRNA-seq - k-means聚类与不同的距离。
Daniel N Baker, Nathan Dyjack, Vladimir Braverman, Stephanie C Hicks, Ben Langmead
Single-cell RNA-sequencing (scRNA-seq) analyses typically begin by clustering a gene-by-cell expression matrix to empirically define groups of cells with similar expression profiles. We describe new methods and a new open source library, minicore, for efficient k-means++ center finding and k-means clustering of scRNA-seq data. Minicore works with sparse count data, as it emerges from typical scRNA-seq experiments, as well as with dense data from after dimensionality reduction. Minicore's novel vectorized weighted reservoir sampling algorithm allows it to find initial k-means++ centers for a 4-million cell dataset in 1.5 minutes using 20 threads. Minicore can cluster using Euclidean distance, but also supports a wider class of measures like Jensen-Shannon Divergence, Kullback-Leibler Divergence, and the Bhattacharyya distance, which can be directly applied to count data and probability distributions. Further, minicore produces lower-cost centerings more efficiently than scikit-learn for scRNA-seq datasets with millions of cells. With careful handling of priors, minicore implements these distance measures with only minor (<2-fold) speed differences among all distances. We show that a minicore pipeline consisting of k-means++, localsearch++ and mini-batch k-means can cluster a 4-million cell dataset in minutes, using less than 10GiB of RAM. This memory-efficiency enables atlas-scale clustering on laptops and other commodity hardware. Finally, we report findings on which distance measures give clusterings that are most consistent with known cell type labels. Availability: The open source library is at https://github.com/dnbaker/minicore. Code used for experiments is at https://github.com/dnbaker/minicore-experiments.
单细胞RNA测序(scRNA-seq)分析通常从逐个细胞的基因表达矩阵聚类开始,以经验定义具有相似表达谱的细胞组。我们描述了用于scRNA-seq数据的有效k-means++中心查找和k-means聚类的新方法和新的开源库minicore。Minicore处理稀疏计数数据,因为它来自典型的scRNA-seq实验,以及降维后的密集数据。Minicore新颖的矢量化加权储层采样算法使其能够使用20个线程在1.5分钟内找到400万个单元数据集的初始k均值++中心。Minicore可以使用欧几里得距离进行聚类,但也支持更广泛的度量,如Jensen Shannon散度、Kullback Leibler散度和Bhattachaiya距离,这些度量可以直接应用于计数数据和概率分布。此外,对于具有数百万个细胞的scRNA-seq数据集,minicore比scikit learn更有效地产生成本更低的中心。通过仔细处理先验,minicore只需少量即可实现这些距离测量(k-means++、localsearch++和迷你批处理k-means可以在几分钟内对400万个细胞数据集进行聚类,使用不到10GiB的RAM。这种内存效率可以在笔记本电脑和其他商品硬件上实现图谱规模的聚类。最后,我们报告了距离测量得出的聚类与已知细胞类型标签最一致的结果。
{"title":"Fast and memory-efficient scRNA-seq <i>k</i>-means clustering with various distances.","authors":"Daniel N Baker,&nbsp;Nathan Dyjack,&nbsp;Vladimir Braverman,&nbsp;Stephanie C Hicks,&nbsp;Ben Langmead","doi":"10.1145/3459930.3469523","DOIUrl":"10.1145/3459930.3469523","url":null,"abstract":"Single-cell RNA-sequencing (scRNA-seq) analyses typically begin by clustering a gene-by-cell expression matrix to empirically define groups of cells with similar expression profiles. We describe new methods and a new open source library, minicore, for efficient k-means++ center finding and k-means clustering of scRNA-seq data. Minicore works with sparse count data, as it emerges from typical scRNA-seq experiments, as well as with dense data from after dimensionality reduction. Minicore's novel vectorized weighted reservoir sampling algorithm allows it to find initial k-means++ centers for a 4-million cell dataset in 1.5 minutes using 20 threads. Minicore can cluster using Euclidean distance, but also supports a wider class of measures like Jensen-Shannon Divergence, Kullback-Leibler Divergence, and the Bhattacharyya distance, which can be directly applied to count data and probability distributions. Further, minicore produces lower-cost centerings more efficiently than scikit-learn for scRNA-seq datasets with millions of cells. With careful handling of priors, minicore implements these distance measures with only minor (<2-fold) speed differences among all distances. We show that a minicore pipeline consisting of k-means++, localsearch++ and mini-batch k-means can cluster a 4-million cell dataset in minutes, using less than 10GiB of RAM. This memory-efficiency enables atlas-scale clustering on laptops and other commodity hardware. Finally, we report findings on which distance measures give clusterings that are most consistent with known cell type labels. Availability: The open source library is at https://github.com/dnbaker/minicore. Code used for experiments is at https://github.com/dnbaker/minicore-experiments.","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":"2021 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8586878/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39733090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Concurrent Imputation and Prediction on EHR data using Bi-Directional GANs: Bi-GANs for EHR imputation and prediction. 使用双向 GANs 对电子病历数据进行同步估算和预测:用于电子病历估算和预测的双向 GANs。
Mehak Gupta, H Timothy Bunnell, Thao-Ly T Phan, Rahmatollah Beheshti

Working with electronic health records (EHRs) is known to be challenging due to several reasons. These reasons include not having: 1) similar lengths (per visit), 2) the same number of observations (per patient), and 3) complete entries in the available records. These issues hinder the performance of the predictive models created using EHRs. In this paper, we approach these issues by presenting a model for the combined task of imputing and predicting values for the irregularly observed and varying length EHR data with missing entries. Our proposed model (dubbed as Bi-GAN) uses a bidirectional recurrent network in a generative adversarial setting. In this architecture, the generator is a bidirectional recurrent network that receives the EHR data and imputes the existing missing values. The discriminator attempts to discriminate between the actual and the imputed values generated by the generator. Using the input data in its entirety, Bi-GAN learns how to impute missing elements in-between (imputation) or outside of the input time steps (prediction). Our method has three advantages to the state-of-the-art methods in the field: (a) one single model performs both the imputation and prediction tasks; (b) the model can perform predictions using time-series of varying length with missing data; (c) it does not require to know the observation and prediction time window during training and can be used for the predictions with different observation and prediction window lengths, for short- and long-term predictions. We evaluate our model on two large EHR datasets to impute and predict body mass index (BMI) values and show its superior performance in both settings.

众所周知,由于多种原因,使用电子健康记录(EHR)具有挑战性。这些原因包括1)相似的时间长度(每次就诊);2)相同的观察次数(每位患者);3)可用记录中的完整条目。这些问题阻碍了使用电子病历创建的预测模型的性能。在本文中,我们提出了一个模型来解决这些问题,该模型可用于对不规则观察和长度不一且条目缺失的电子病历数据进行估算和预测。我们提出的模型(称为 Bi-GAN)在生成对抗环境中使用双向循环网络。在这一架构中,生成器是一个双向递归网络,它接收电子病历数据并对现有的缺失值进行估算。判别器试图在生成器生成的实际值和估算值之间进行判别。Bi-GAN 使用完整的输入数据,学习如何在输入时间步骤之间(估算)或之外(预测)估算缺失元素。与该领域最先进的方法相比,我们的方法有三个优势:(a) 一个模型就能同时完成估算和预测任务;(b) 该模型可以使用不同长度的时间序列和缺失数据进行预测;(c) 在训练过程中不需要知道观察和预测的时间窗口,可用于不同观察和预测窗口长度的预测,也可用于短期和长期预测。我们在两个大型电子病历数据集上对我们的模型进行了评估,以估算和预测身体质量指数(BMI)值,结果显示该模型在这两种情况下都表现出色。
{"title":"Concurrent Imputation and Prediction on EHR data using Bi-Directional GANs: Bi-GANs for EHR imputation and prediction.","authors":"Mehak Gupta, H Timothy Bunnell, Thao-Ly T Phan, Rahmatollah Beheshti","doi":"10.1145/3459930.3469512","DOIUrl":"10.1145/3459930.3469512","url":null,"abstract":"<p><p>Working with electronic health records (EHRs) is known to be challenging due to several reasons. These reasons include not having: 1) similar lengths (per visit), 2) the same number of observations (per patient), and 3) complete entries in the available records. These issues hinder the performance of the predictive models created using EHRs. In this paper, we approach these issues by presenting a model for the combined task of imputing and predicting values for the irregularly observed and varying length EHR data with missing entries. Our proposed model (dubbed as Bi-GAN) uses a bidirectional recurrent network in a generative adversarial setting. In this architecture, the generator is a bidirectional recurrent network that receives the EHR data and imputes the existing missing values. The discriminator attempts to discriminate between the actual and the imputed values generated by the generator. Using the input data in its entirety, Bi-GAN learns how to impute missing elements in-between (imputation) or outside of the input time steps (prediction). Our method has three advantages to the state-of-the-art methods in the field: (a) one single model performs both the imputation and prediction tasks; (b) the model can perform predictions using time-series of varying length with missing data; (c) it does not require to know the observation and prediction time window during training and can be used for the predictions with different observation and prediction window lengths, for short- and long-term predictions. We evaluate our model on two large EHR datasets to impute and predict body mass index (BMI) values and show its superior performance in both settings.</p>","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":"2021 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8482531/pdf/nihms-1740754.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39483618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Joint Learning for Biomedical NER and Entity Normalization: Encoding Schemes, Counterfactual Examples, and Zero-Shot Evaluation. 生物医学NER和实体归一化的联合学习:编码方案,反事实示例和零射击评估。
Jiho Noh, Ramakanth Kavuluru

Named entity recognition (NER) and normalization (EN) form an indispensable first step to many biomedical natural language processing applications. In biomedical information science, recognizing entities (e.g., genes, diseases, or drugs) and normalizing them to concepts in standard terminologies or thesauri (e.g., Entrez, ICD-10, or RxNorm) is crucial for identifying more informative relations among them that drive disease etiology, progression, and treatment. In this effort we pursue two high level strategies to improve biomedical ER and EN. The first is to decouple standard entity encoding tags (e.g., "B-Drug" for the beginning of a drug) into type tags (e.g., "Drug") and positional tags (e.g., "B"). A second strategy is to use additional counterfactual training examples to handle the issue of models learning spurious correlations between surrounding context and normalized concepts in training data. We conduct elaborate experiments using the MedMentions dataset, the largest dataset of its kind for ER and EN in biomedicine. We find that our first strategy performs better in entity normalization when compared with the standard coding scheme. The second data augmentation strategy uniformly improves performance in span detection, typing, and normalization. The gains from counterfactual examples are more prominent when evaluating in zero-shot settings, for concepts that have never been encountered during training.

命名实体识别(NER)和归一化(EN)是许多生物医学自然语言处理应用不可或缺的第一步。在生物医学信息科学中,识别实体(如基因、疾病或药物)并将其规范化为标准术语或词典中的概念(如Entrez、ICD-10或RxNorm)对于确定它们之间驱动疾病病因、进展和治疗的更多信息关系至关重要。在这项工作中,我们追求两个高水平的战略,以提高生物医学ER和EN。首先是将标准实体编码标签(例如,“B- drug”表示药物的开头)解耦为类型标签(例如,“drug”)和位置标签(例如,“B”)。第二种策略是使用额外的反事实训练示例来处理模型在训练数据中学习周围上下文和规范化概念之间的虚假关联的问题。我们使用med提及数据集进行了详细的实验,med提及数据集是生物医学中同类最大的ER和EN数据集。我们发现,与标准编码方案相比,我们的第一种策略在实体规范化方面表现更好。第二种数据增强策略统一地提高了跨度检测、类型和规范化方面的性能。当在零射击设置中评估时,对于训练中从未遇到过的概念,反事实示例的收益更加突出。
{"title":"Joint Learning for Biomedical NER and Entity Normalization: Encoding Schemes, Counterfactual Examples, and Zero-Shot Evaluation.","authors":"Jiho Noh,&nbsp;Ramakanth Kavuluru","doi":"10.1145/3459930.3469533","DOIUrl":"https://doi.org/10.1145/3459930.3469533","url":null,"abstract":"<p><p>Named entity recognition (NER) and normalization (EN) form an indispensable first step to many biomedical natural language processing applications. In biomedical information science, recognizing entities (e.g., genes, diseases, or drugs) and normalizing them to concepts in standard terminologies or thesauri (e.g., Entrez, ICD-10, or RxNorm) is crucial for identifying more informative relations among them that drive disease etiology, progression, and treatment. In this effort we pursue two high level strategies to improve biomedical ER and EN. The first is to decouple standard entity encoding tags (e.g., \"B-Drug\" for the beginning of a drug) into type tags (e.g., \"Drug\") and positional tags (e.g., \"B\"). A second strategy is to use additional counterfactual training examples to handle the issue of models learning spurious correlations between surrounding context and normalized concepts in training data. We conduct elaborate experiments using the MedMentions dataset, the largest dataset of its kind for ER and EN in biomedicine. We find that our first strategy performs better in entity normalization when compared with the standard coding scheme. The second data augmentation strategy uniformly improves performance in span detection, typing, and normalization. The gains from counterfactual examples are more prominent when evaluating in zero-shot settings, for concepts that have never been encountered during training.</p>","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":"2021 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1145/3459930.3469533","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39402820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Transformer-Based Named Entity Recognition for Parsing Clinical Trial Eligibility Criteria. 基于变压器的命名实体识别分析临床试验资格标准。
Shubo Tian, Arslan Erdengasileng, Xi Yang, Yi Guo, Yonghui Wu, Jinfeng Zhang, Jiang Bian, Zhe He

The rapid adoption of electronic health records (EHRs) systems has made clinical data available in electronic format for research and for many downstream applications. Electronic screening of potentially eligible patients using these clinical databases for clinical trials is a critical need to improve trial recruitment efficiency. Nevertheless, manually translating free-text eligibility criteria into database queries is labor intensive and inefficient. To facilitate automated screening, free-text eligibility criteria must be structured and coded into a computable format using controlled vocabularies. Named entity recognition (NER) is thus an important first step. In this study, we evaluate 4 state-of-the-art transformer-based NER models on two publicly available annotated corpora of eligibility criteria released by Columbia University (i.e., the Chia data) and Facebook Research (i.e.the FRD data). Four transformer-based models (i.e., BERT, ALBERT, RoBERTa, and ELECTRA) pretrained with general English domain corpora vs. those pretrained with PubMed citations, clinical notes from the MIMIC-III dataset and eligibility criteria extracted from all the clinical trials on ClinicalTrials.gov were compared. Experimental results show that RoBERTa pretrained with MIMIC-III clinical notes and eligibility criteria yielded the highest strict and relaxed F-scores in both the Chia data (i.e., 0.658/0.798) and the FRD data (i.e., 0.785/0.916). With promising NER results, further investigations on building a reliable natural language processing (NLP)-assisted pipeline for automated electronic screening are needed.

电子健康记录(EHRs)系统的迅速采用使得临床数据以电子格式可用于研究和许多下游应用。使用这些临床数据库进行临床试验的潜在合格患者的电子筛选是提高试验招募效率的关键需要。然而,手动将自由文本资格标准转换为数据库查询是劳动密集型且效率低下的。为了方便自动筛选,必须使用受控词汇表将自由文本资格标准结构化并编码为可计算的格式。命名实体识别(NER)因此是重要的第一步。在本研究中,我们在哥伦比亚大学(即中国数据)和Facebook Research(即FRD数据)发布的两个公开可用的资格标准注释语料库上评估了4个最先进的基于变压器的NER模型。四种基于转换器的模型(即BERT、ALBERT、RoBERTa和ELECTRA)使用通用英语领域语料库进行预训练,与使用PubMed引文、MIMIC-III数据集的临床记录和从ClinicalTrials.gov上提取的所有临床试验的资格标准进行预训练的模型进行比较。实验结果表明,使用MIMIC-III临床记录和资格标准进行预训练的RoBERTa在Chia数据(0.658/0.798)和FRD数据(0.785/0.916)中均获得了最高的严格f分和宽松f分。随着NER结果的出现,需要进一步研究建立一个可靠的自然语言处理(NLP)辅助的自动化电子筛选管道。
{"title":"Transformer-Based Named Entity Recognition for Parsing Clinical Trial Eligibility Criteria.","authors":"Shubo Tian,&nbsp;Arslan Erdengasileng,&nbsp;Xi Yang,&nbsp;Yi Guo,&nbsp;Yonghui Wu,&nbsp;Jinfeng Zhang,&nbsp;Jiang Bian,&nbsp;Zhe He","doi":"10.1145/3459930.3469560","DOIUrl":"https://doi.org/10.1145/3459930.3469560","url":null,"abstract":"<p><p>The rapid adoption of electronic health records (EHRs) systems has made clinical data available in electronic format for research and for many downstream applications. Electronic screening of potentially eligible patients using these clinical databases for clinical trials is a critical need to improve trial recruitment efficiency. Nevertheless, manually translating free-text eligibility criteria into database queries is labor intensive and inefficient. To facilitate automated screening, free-text eligibility criteria must be structured and coded into a computable format using controlled vocabularies. Named entity recognition (NER) is thus an important first step. In this study, we evaluate 4 state-of-the-art transformer-based NER models on two publicly available annotated corpora of eligibility criteria released by Columbia University (i.e., the Chia data) and Facebook Research (i.e.the FRD data). Four transformer-based models (i.e., BERT, ALBERT, RoBERTa, and ELECTRA) pretrained with general English domain corpora vs. those pretrained with PubMed citations, clinical notes from the MIMIC-III dataset and eligibility criteria extracted from all the clinical trials on ClinicalTrials.gov were compared. Experimental results show that RoBERTa pretrained with MIMIC-III clinical notes and eligibility criteria yielded the highest strict and relaxed F-scores in both the Chia data (i.e., 0.658/0.798) and the FRD data (i.e., 0.785/0.916). With promising NER results, further investigations on building a reliable natural language processing (NLP)-assisted pipeline for automated electronic screening are needed.</p>","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":"2021 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1145/3459930.3469560","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39328500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
KGDAL: Knowledge Graph Guided Double Attention LSTM for Rolling Mortality Prediction for AKI-D Patients. KGDAL:知识图谱引导双注意LSTM用于AKI-D患者滚动死亡率预测。
Lucas Jing Liu, Victor Ortiz-Soriano, Javier A Neyra, Jin Chen

With the rapid accumulation of electronic health record (EHR) data, deep learning (DL) models have exhibited promising performance on patient risk prediction. Recent advances have also demonstrated the effectiveness of knowledge graphs (KG) in providing valuable prior knowledge for further improving DL model performance. However, it is still unclear how KG can be utilized to encode high-order relations among clinical concepts and how DL models can make full use of the encoded concept relations to solve real-world healthcare problems and to interpret the outcomes. We propose a novel knowledge graph guided double attention LSTM model named KGDAL for rolling mortality prediction for critically ill patients with acute kidney injury requiring dialysis (AKI-D). KGDAL constructs a KG-based two-dimension attention in both time and feature spaces. In the experiment with two large healthcare datasets, we compared KGDAL with a variety of rolling mortality prediction models and conducted an ablation study to test the effectiveness, efficacy, and contribution of different attention mechanisms. The results showed that KGDAL clearly outperformed all the compared models. Also, KGDAL-derived patient risk trajectories may assist healthcare providers to make timely decisions and actions. The source code, sample data, and manual of KGDAL are available at https://github.com/lucasliu0928/KGDAL.

随着电子病历(EHR)数据的快速积累,深度学习(DL)模型在患者风险预测方面表现出了良好的性能。最近的进展也证明了知识图(KG)在为进一步提高深度学习模型性能提供有价值的先验知识方面的有效性。然而,KG如何用于编码临床概念之间的高阶关系,以及DL模型如何充分利用编码的概念关系来解决现实世界的医疗问题并解释结果,目前尚不清楚。我们提出了一种新的知识图谱引导的双注意LSTM模型KGDAL,用于预测急性肾损伤需要透析的危重患者(AKI-D)的滚动死亡率。KGDAL在时间和特征空间上构建了基于kg的二维注意力。在两个大型医疗数据集的实验中,我们将KGDAL与各种滚动死亡率预测模型进行了比较,并进行了消融研究,以测试不同注意机制的有效性、疗效和贡献。结果表明,KGDAL明显优于所有比较模型。此外,kgdal衍生的患者风险轨迹可以帮助医疗保健提供者及时做出决策和采取行动。KGDAL的源代码、样例数据和手册可在https://github.com/lucasliu0928/KGDAL上获得。
{"title":"KGDAL: Knowledge Graph Guided Double Attention LSTM for Rolling Mortality Prediction for AKI-D Patients.","authors":"Lucas Jing Liu, Victor Ortiz-Soriano, Javier A Neyra, Jin Chen","doi":"10.1145/3459930.3469513","DOIUrl":"10.1145/3459930.3469513","url":null,"abstract":"<p><p>With the rapid accumulation of electronic health record (EHR) data, deep learning (DL) models have exhibited promising performance on patient risk prediction. Recent advances have also demonstrated the effectiveness of knowledge graphs (KG) in providing valuable prior knowledge for further improving DL model performance. However, it is still unclear how KG can be utilized to encode high-order relations among clinical concepts and how DL models can make full use of the encoded concept relations to solve real-world healthcare problems and to interpret the outcomes. We propose a novel knowledge graph guided double attention LSTM model named KGDAL for rolling mortality prediction for critically ill patients with acute kidney injury requiring dialysis (AKI-D). KGDAL constructs a KG-based two-dimension attention in both time and feature spaces. In the experiment with two large healthcare datasets, we compared KGDAL with a variety of rolling mortality prediction models and conducted an ablation study to test the effectiveness, efficacy, and contribution of different attention mechanisms. The results showed that KGDAL clearly outperformed all the compared models. Also, KGDAL-derived patient risk trajectories may assist healthcare providers to make timely decisions and actions. The source code, sample data, and manual of KGDAL are available at https://github.com/lucasliu0928/KGDAL.</p>","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":"2021 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8445228/pdf/nihms-1737960.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39453029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised manifold alignment for single-cell multi-omics data. 单细胞多组学数据的无监督流形对齐。
Ritambhara Singh, Pinar Demetci, Giancarlo Bonora, Vijay Ramani, Choli Lee, He Fang, Zhijun Duan, Xinxian Deng, Jay Shendure, Christine Disteche, William Stafford Noble

Integrating single-cell measurements that capture different properties of the genome is vital to extending our understanding of genome biology. This task is challenging due to the lack of a shared axis across datasets obtained from different types of single-cell experiments. For most such datasets, we lack corresponding information among the cells (samples) and the measurements (features). In this scenario, unsupervised algorithms that are capable of aligning single-cell experiments are critical to learning an in silico co-assay that can help draw correspondences among the cells. Maximum mean discrepancy-based manifold alignment (MMD-MA) is such an unsupervised algorithm. Without requiring correspondence information, it can align single-cell datasets from different modalities in a common shared latent space, showing promising results on simulations and a small-scale single-cell experiment with 61 cells. However, it is essential to explore the applicability of this method to larger single-cell experiments with thousands of cells so that it can be of practical interest to the community. In this paper, we apply MMD-MA to two recent datasets that measure transcriptome and chromatin accessibility in ~2000 single cells. To scale the runtime of MMD-MA to a more substantial number of cells, we extend the original implementation to run on GPUs. We also introduce a method to automatically select one of the user-defined parameters, thus reducing the hyperparameter search space. We demonstrate that the proposed extensions allow MMD-MA to accurately align state-of-the-art single-cell experiments.

整合捕获基因组不同特性的单细胞测量对于扩展我们对基因组生物学的理解至关重要。由于从不同类型的单细胞实验中获得的数据集之间缺乏共享轴,因此这项任务具有挑战性。对于大多数这样的数据集,我们缺乏单元(样本)和测量(特征)之间的相应信息。在这种情况下,能够对齐单细胞实验的无监督算法对于学习可以帮助绘制细胞之间对应关系的计算机联合分析至关重要。基于最大平均误差的流形对齐(MMD-MA)就是这样一种无监督算法。在不需要对应信息的情况下,它可以将来自不同模态的单细胞数据集对齐在一个共同的潜在空间中,在模拟和61个细胞的小规模单细胞实验中显示出令人鼓舞的结果。然而,探索这种方法在数千个细胞的大型单细胞实验中的适用性是至关重要的,这样它才能对社区产生实际的兴趣。在本文中,我们将MMD-MA应用于两个最近的数据集,这些数据集测量了约2000个单细胞的转录组和染色质可及性。为了将MMD-MA的运行时扩展到更大数量的单元,我们扩展了原始实现以在gpu上运行。我们还引入了一种自动选择用户自定义参数的方法,从而减少了超参数搜索空间。我们证明,提出的扩展允许MMD-MA准确地对准最先进的单细胞实验。
{"title":"Unsupervised manifold alignment for single-cell multi-omics data.","authors":"Ritambhara Singh,&nbsp;Pinar Demetci,&nbsp;Giancarlo Bonora,&nbsp;Vijay Ramani,&nbsp;Choli Lee,&nbsp;He Fang,&nbsp;Zhijun Duan,&nbsp;Xinxian Deng,&nbsp;Jay Shendure,&nbsp;Christine Disteche,&nbsp;William Stafford Noble","doi":"10.1145/3388440.3412410","DOIUrl":"https://doi.org/10.1145/3388440.3412410","url":null,"abstract":"<p><p>Integrating single-cell measurements that capture different properties of the genome is vital to extending our understanding of genome biology. This task is challenging due to the lack of a shared axis across datasets obtained from different types of single-cell experiments. For most such datasets, we lack corresponding information among the cells (samples) and the measurements (features). In this scenario, unsupervised algorithms that are capable of aligning single-cell experiments are critical to learning an <i>in silico</i> co-assay that can help draw correspondences among the cells. Maximum mean discrepancy-based manifold alignment (MMD-MA) is such an unsupervised algorithm. Without requiring correspondence information, it can align single-cell datasets from different modalities in a common shared latent space, showing promising results on simulations and a small-scale single-cell experiment with 61 cells. However, it is essential to explore the applicability of this method to larger single-cell experiments with thousands of cells so that it can be of practical interest to the community. In this paper, we apply MMD-MA to two recent datasets that measure transcriptome and chromatin accessibility in ~2000 single cells. To scale the runtime of MMD-MA to a more substantial number of cells, we extend the original implementation to run on GPUs. We also introduce a method to automatically select one of the user-defined parameters, thus reducing the hyperparameter search space. We demonstrate that the proposed extensions allow MMD-MA to accurately align state-of-the-art single-cell experiments.</p>","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":"2020 ","pages":"1-10"},"PeriodicalIF":0.0,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1145/3388440.3412410","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10130200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
A deep learning fusion model for brain disorder classification: Application to distinguishing schizophrenia and autism spectrum disorder. 用于脑部疾病分类的深度学习融合模型:应用于区分精神分裂症和自闭症谱系障碍。
Yuhui Du, Bang Li, Yuliang Hou, Vince D Calhoun

Deep learning has shown a great promise in classifying brain disorders due to its powerful ability in learning optimal features by nonlinear transformation. However, given the high-dimension property of neuroimaging data, how to jointly exploit complementary information from multimodal neuroimaging data in deep learning is difficult. In this paper, we propose a novel multilevel convolutional neural network (CNN) fusion method that can effectively combine different types of neuroimage-derived features. Importantly, we incorporate a sequential feature selection into the CNN model to increase the feature interpretability. To evaluate our method, we classified two symptom-related brain disorders using large-sample multi-site data from 335 schizophrenia (SZ) patients and 380 autism spectrum disorder (ASD) patients within a cross-validation procedure. Brain functional networks, functional network connectivity, and brain structural morphology were employed to provide possible features. As expected, our fusion method outperformed the CNN model using only single type of features, as our method yielded higher classification accuracy (with mean accuracy >85%) and was more reliable across multiple runs in differentiating the two groups. We found that the default mode, cognitive control, and subcortical regions contributed more in their distinction. Taken together, our method provides an effective means to fuse multimodal features for the diagnosis of different psychiatric and neurological disorders.

深度学习通过非线性变换学习最优特征的强大能力,使其在脑疾病分类方面大有可为。然而,鉴于神经影像数据的高维特性,如何在深度学习中联合利用多模态神经影像数据的互补信息是一个难题。在本文中,我们提出了一种新颖的多级卷积神经网络(CNN)融合方法,它能有效地结合不同类型的神经图像特征。重要的是,我们在 CNN 模型中加入了顺序特征选择,以提高特征的可解释性。为了评估我们的方法,我们使用来自 335 名精神分裂症(SZ)患者和 380 名自闭症谱系障碍(ASD)患者的大样本多站点数据,在交叉验证程序中对两种症状相关的脑部疾病进行了分类。大脑功能网络、功能网络连通性和大脑结构形态被用来提供可能的特征。不出所料,我们的融合方法优于只使用单一类型特征的 CNN 模型,因为我们的方法获得了更高的分类准确率(平均准确率大于 85%),并且在多次运行中区分两组的可靠性更高。我们发现,默认模式、认知控制和皮层下区域对它们的区分贡献更大。综上所述,我们的方法为融合多模态特征诊断不同的精神和神经疾病提供了有效手段。
{"title":"A deep learning fusion model for brain disorder classification: Application to distinguishing schizophrenia and autism spectrum disorder.","authors":"Yuhui Du, Bang Li, Yuliang Hou, Vince D Calhoun","doi":"10.1145/3388440.3412478","DOIUrl":"10.1145/3388440.3412478","url":null,"abstract":"<p><p>Deep learning has shown a great promise in classifying brain disorders due to its powerful ability in learning optimal features by nonlinear transformation. However, given the high-dimension property of neuroimaging data, how to jointly exploit complementary information from multimodal neuroimaging data in deep learning is difficult. In this paper, we propose a novel multilevel convolutional neural network (CNN) fusion method that can effectively combine different types of neuroimage-derived features. Importantly, we incorporate a sequential feature selection into the CNN model to increase the feature interpretability. To evaluate our method, we classified two symptom-related brain disorders using large-sample multi-site data from 335 schizophrenia (SZ) patients and 380 autism spectrum disorder (ASD) patients within a cross-validation procedure. Brain functional networks, functional network connectivity, and brain structural morphology were employed to provide possible features. As expected, our fusion method outperformed the CNN model using only single type of features, as our method yielded higher classification accuracy (with mean accuracy >85%) and was more reliable across multiple runs in differentiating the two groups. We found that the default mode, cognitive control, and subcortical regions contributed more in their distinction. Taken together, our method provides an effective means to fuse multimodal features for the diagnosis of different psychiatric and neurological disorders.</p>","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7758676/pdf/nihms-1654686.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38750792","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Combine Cryo-EM Density Map and Residue Contact for Protein Structure Prediction - A Case Study. 结合低温电镜密度图和残馀接触蛋白结构预测-一个案例研究。
Maytha Alshammari, Jing He

Cryo-electron microscopy is a major structure determination technique for large molecular machines and membrane-associated complexes. Although atomic structures have been determined directly from cryo-EM density maps with high resolutions, current structure determination methods for medium resolution (5 to 10 Å) cryo-EM maps are limited by the availability of structure templates. Secondary structure traces are lines detected from a cryo-EM density map for α-helices and β-strands of a protein. When combined with secondary structure sequence segments predicted from a protein sequence, it is possible to generate a set of likely topologies of α-traces and β-sheet traces. A topology describes the overall folding relationship among secondary structures; it is a critical piece of information for deriving the corresponding atomic structure. We propose a method for protein structure prediction that combines three sources of information: the secondary structure traces detected from the cryo-EM density map, predicted secondary structure sequence segments, and amino acid contact pairs predicted using MULTICOM. A case study shows that using amino acid contact prediction from MULTICOM improves the ranking of the true topology. Our observations convey that using a small set of highly voted secondary structure contact pairs enhances the ranking in all experiments conducted for this case.

低温电子显微镜是大分子机器和膜相关复合物的主要结构测定技术。虽然原子结构已经直接从高分辨率的低温电镜密度图中确定,但目前用于中分辨率(5到10 Å)低温电镜图的结构确定方法受到结构模板可用性的限制。二级结构痕迹是从蛋白质的α-螺旋和β-链的低温电镜密度图中检测到的线。当结合从蛋白质序列中预测的二级结构序列片段时,可以生成一组α-示踪和β-示踪的可能拓扑结构。拓扑描述二级结构之间的整体折叠关系;这是推导相应原子结构的关键信息。我们提出了一种结合三种信息来源的蛋白质结构预测方法:从低温电镜密度图中检测到的二级结构痕迹,预测的二级结构序列片段,以及使用MULTICOM预测的氨基酸接触对。实例研究表明,利用MULTICOM的氨基酸接触预测提高了真实拓扑的排序。我们的观察表明,使用一组高度投票的二级结构接触对提高了在这种情况下进行的所有实验中的排名。
{"title":"Combine Cryo-EM Density Map and Residue Contact for Protein Structure Prediction - A Case Study.","authors":"Maytha Alshammari,&nbsp;Jing He","doi":"10.1145/3388440.3414708","DOIUrl":"https://doi.org/10.1145/3388440.3414708","url":null,"abstract":"<p><p>Cryo-electron microscopy is a major structure determination technique for large molecular machines and membrane-associated complexes. Although atomic structures have been determined directly from cryo-EM density maps with high resolutions, current structure determination methods for medium resolution (5 to 10 Å) cryo-EM maps are limited by the availability of structure templates. Secondary structure traces are lines detected from a cryo-EM density map for α-helices and β-strands of a protein. When combined with secondary structure sequence segments predicted from a protein sequence, it is possible to generate a set of likely topologies of α-traces and β-sheet traces. A topology describes the overall folding relationship among secondary structures; it is a critical piece of information for deriving the corresponding atomic structure. We propose a method for protein structure prediction that combines three sources of information: the secondary structure traces detected from the cryo-EM density map, predicted secondary structure sequence segments, and amino acid contact pairs predicted using MULTICOM. A case study shows that using amino acid contact prediction from MULTICOM improves the ranking of the true topology. Our observations convey that using a small set of highly voted secondary structure contact pairs enhances the ranking in all experiments conducted for this case.</p>","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1145/3388440.3414708","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40524905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Using Curriculum Learning in Pattern Recognition of 3-dimensional Cryo-electron Microscopy Density Maps. 课程学习在三维冷冻电镜密度图模式识别中的应用。
Yangmei Deng, Yongcheng Mu, Salim Sazzed, Jiangwen Sun, Jing He

Although Cryo-electron microscopy (cryo-EM) has been successfully used to derive atomic structures for many proteins, it is still challenging to derive atomic structure when the resolution of cryo-EM density maps is in the medium range, e.g., 5-10 Å. Studies have attempted to utilize machine learning methods, especially deep neural networks to build predictive models for the detection of protein secondary structures from cryo-EM images, which ultimately helps to derive the atomic structure of proteins. However, the large variation in data quality makes it challenging to train a deep neural network with high prediction accuracy. Curriculum learning has been shown as an effective learning paradigm in machine learning. In this paper, we present a study using curriculum learning as a more effective way to utilize cryo-EM density maps with varying quality. We investigated three distinct training curricula that differ in whether/how images used for training in past are reused while the network was continually trained using new images. A total of 1,382 3-dimensional cryo-EM images were extracted from density maps of Electron Microscopy Data Bank in our study. Our results indicate learning with curriculum significantly improves the performance of the final trained network when the forgetting problem is properly addressed.

尽管冷冻电子显微镜(cryo-EM)已经成功地用于推导许多蛋白质的原子结构,但当冷冻电子显微镜密度图的分辨率在中等范围内(例如5-10 Å)时,推导原子结构仍然具有挑战性。研究试图利用机器学习方法,特别是深度神经网络来建立预测模型,用于从冷冻电镜图像中检测蛋白质二级结构,最终有助于推导蛋白质的原子结构。然而,数据质量的巨大变化给训练具有高预测精度的深度神经网络带来了挑战。课程学习已被证明是机器学习中一种有效的学习范式。在本文中,我们提出了一项研究,使用课程学习作为一种更有效的方法来利用不同质量的低温电镜密度图。我们研究了三种不同的训练课程,它们在使用新图像不断训练网络的同时,是否/如何重复使用过去用于训练的图像。本研究从电子显微镜数据库的密度图中提取了1382张三维冷冻电镜图像。我们的研究结果表明,当遗忘问题得到适当解决时,课程学习显著提高了最终训练网络的性能。
{"title":"Using Curriculum Learning in Pattern Recognition of 3-dimensional Cryo-electron Microscopy Density Maps.","authors":"Yangmei Deng,&nbsp;Yongcheng Mu,&nbsp;Salim Sazzed,&nbsp;Jiangwen Sun,&nbsp;Jing He","doi":"10.1145/3388440.3414710","DOIUrl":"https://doi.org/10.1145/3388440.3414710","url":null,"abstract":"<p><p>Although Cryo-electron microscopy (cryo-EM) has been successfully used to derive atomic structures for many proteins, it is still challenging to derive atomic structure when the resolution of cryo-EM density maps is in the medium range, e.g., 5-10 Å. Studies have attempted to utilize machine learning methods, especially deep neural networks to build predictive models for the detection of protein secondary structures from cryo-EM images, which ultimately helps to derive the atomic structure of proteins. However, the large variation in data quality makes it challenging to train a deep neural network with high prediction accuracy. Curriculum learning has been shown as an effective learning paradigm in machine learning. In this paper, we present a study using curriculum learning as a more effective way to utilize cryo-EM density maps with varying quality. We investigated three distinct training curricula that differ in whether/how images used for training in past are reused while the network was continually trained using new images. A total of 1,382 3-dimensional cryo-EM images were extracted from density maps of Electron Microscopy Data Bank in our study. Our results indicate learning with curriculum significantly improves the performance of the final trained network when the forgetting problem is properly addressed.</p>","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1145/3388440.3414710","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40507888","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Correlation Imputation in Single cell RNA-seq using Auxiliary Information and Ensemble Learning. 基于辅助信息和集成学习的单细胞RNA-seq相关归算。
Luqin Gan, Giuseppe Vinci, Genevera I Allen

Single cell RNA sequencing is a powerful technique that measures the gene expression of individual cells in a high throughput fashion. However, due to sequencing inefficiency, the data is unreliable due to dropout events, or technical artifacts where genes erroneously appear to have zero expression. Many data imputation methods have been proposed to alleviate this issue. Yet, effective imputation can be difficult and biased because the data is sparse and high-dimensional, resulting in major distortions in downstream analyses. In this paper, we propose a completely novel approach that imputes the gene-by-gene correlations rather than the data itself. We call this method SCENA: Single cell RNA-seq Correlation completion by ENsemble learning and Auxiliary information. The SCENA gene-by-gene correlation matrix estimate is obtained by model stacking of multiple imputed correlation matrices based on known auxiliary information about gene connections. In an extensive simulation study based on real scRNA-seq data, we demonstrate that SCENA not only accurately imputes gene correlations but also outperforms existing imputation approaches in downstream analyses such as dimension reduction, cell clustering, graphical model estimation.

单细胞RNA测序是一种强大的技术,以高通量的方式测量单个细胞的基因表达。然而,由于测序效率低下,数据是不可靠的,由于辍学事件,或技术工件,基因错误地表现为零表达。为了缓解这一问题,人们提出了许多数据输入方法。然而,由于数据稀疏和高维,有效的归算可能是困难的和有偏差的,导致下游分析中的主要扭曲。在本文中,我们提出了一种全新的方法,计算基因间的相关性,而不是数据本身。我们称这种方法为SCENA:单细胞RNA-seq相关的集成学习和辅助信息完成。基于已知的基因连接辅助信息,对多个输入的相关矩阵进行模型叠加,得到SCENA基因间的相关矩阵估计。在一项基于真实scRNA-seq数据的广泛模拟研究中,我们证明了SCENA不仅准确地推测基因相关性,而且在下游分析(如降维、细胞聚类、图形模型估计)中优于现有的推测方法。
{"title":"Correlation Imputation in Single cell RNA-seq using Auxiliary Information and Ensemble Learning.","authors":"Luqin Gan,&nbsp;Giuseppe Vinci,&nbsp;Genevera I Allen","doi":"10.1145/3388440.3412462","DOIUrl":"https://doi.org/10.1145/3388440.3412462","url":null,"abstract":"<p><p>Single cell RNA sequencing is a powerful technique that measures the gene expression of individual cells in a high throughput fashion. However, due to sequencing inefficiency, the data is unreliable due to dropout events, or technical artifacts where genes erroneously appear to have zero expression. Many data imputation methods have been proposed to alleviate this issue. Yet, effective imputation can be difficult and biased because the data is sparse and high-dimensional, resulting in major distortions in downstream analyses. In this paper, we propose a completely novel approach that imputes the gene-by-gene correlations rather than the data itself. We call this method SCENA: Single cell RNA-seq Correlation completion by ENsemble learning and Auxiliary information. The SCENA gene-by-gene correlation matrix estimate is obtained by model stacking of multiple imputed correlation matrices based on known auxiliary information about gene connections. In an extensive simulation study based on real scRNA-seq data, we demonstrate that SCENA not only accurately imputes gene correlations but also outperforms existing imputation approaches in downstream analyses such as dimension reduction, cell clustering, graphical model estimation.</p>","PeriodicalId":72044,"journal":{"name":"ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1145/3388440.3412462","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39197526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
ACM-BCB ... ... : the ... ACM Conference on Bioinformatics, Computational Biology and Biomedicine. ACM Conference on Bioinformatics, Computational Biology and Biomedicine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1