DSAGLSTM-DTA: Prediction of Drug-Target Affinity using Dual Self-Attention and LSTM

Lyu Zhijian, Shaohua Jiang, Yonghao Tan
{"title":"DSAGLSTM-DTA: Prediction of Drug-Target Affinity using Dual Self-Attention and LSTM","authors":"Lyu Zhijian, Shaohua Jiang, Yonghao Tan","doi":"10.5121/mlaij.2022.9201","DOIUrl":null,"url":null,"abstract":"The research on affinity between drugs and targets (DTA) aims to effectively narrow the target search space for drug repurposing. Therefore, reasonable prediction of drug and target affinities can minimize the waste of resources such as human and material resources. In this work, a novel graph-based model called DSAGLSTM-DTA was proposed for DTA prediction. The proposed model is unlike previous graph-based drug-target affinity model, which incorporated self-attention mechanisms in the feature extraction process of drug molecular graphs to fully extract its effective feature representations. The features of each atom in the 2D molecular graph were weighted based on attention score before being aggregated as molecule representation and two distinct pooling architectures, namely centralized and distributed architectures were implemented and compared on benchmark datasets. In addition, in the course of processing protein sequences, inspired by the approach of protein feature extraction in GDGRU-DTA, we continue to interpret protein sequences as time series and extract their features using Bidirectional Long Short-Term Memory (BiLSTM) networks, since the context-dependence of long amino acid sequences. Similarly, DSAGLSTM-DTA also utilized a self-attention mechanism in the process of protein feature extraction to obtain comprehensive representations of proteins, in which the final hidden states for element in the batch were weighted with the each unit output of LSTM, and the results were represented as the final feature of proteins. Eventually, representations of drug and protein were concatenated and fed into prediction block for final prediction. The proposed model was evaluated on different regression datasets and binary classification datasets, and the results demonstrated that DSAGLSTM-DTA was superior to some state-ofthe-art DTA models and exhibited good generalization ability.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Machine Learning and Applications: An International Journal","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5121/mlaij.2022.9201","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The research on affinity between drugs and targets (DTA) aims to effectively narrow the target search space for drug repurposing. Therefore, reasonable prediction of drug and target affinities can minimize the waste of resources such as human and material resources. In this work, a novel graph-based model called DSAGLSTM-DTA was proposed for DTA prediction. The proposed model is unlike previous graph-based drug-target affinity model, which incorporated self-attention mechanisms in the feature extraction process of drug molecular graphs to fully extract its effective feature representations. The features of each atom in the 2D molecular graph were weighted based on attention score before being aggregated as molecule representation and two distinct pooling architectures, namely centralized and distributed architectures were implemented and compared on benchmark datasets. In addition, in the course of processing protein sequences, inspired by the approach of protein feature extraction in GDGRU-DTA, we continue to interpret protein sequences as time series and extract their features using Bidirectional Long Short-Term Memory (BiLSTM) networks, since the context-dependence of long amino acid sequences. Similarly, DSAGLSTM-DTA also utilized a self-attention mechanism in the process of protein feature extraction to obtain comprehensive representations of proteins, in which the final hidden states for element in the batch were weighted with the each unit output of LSTM, and the results were represented as the final feature of proteins. Eventually, representations of drug and protein were concatenated and fed into prediction block for final prediction. The proposed model was evaluated on different regression datasets and binary classification datasets, and the results demonstrated that DSAGLSTM-DTA was superior to some state-ofthe-art DTA models and exhibited good generalization ability.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
DSAGLSTM-DTA:利用双自注意和LSTM预测药物-靶标亲和力
药物与靶点亲和力(DTA)的研究旨在有效缩小药物再利用的靶点搜索空间。因此,合理预测药物和靶标亲和力,可以最大限度地减少人力、物力等资源的浪费。在这项工作中,提出了一种新的基于图的DTA预测模型DSAGLSTM-DTA。该模型不同于以往基于图的药物-靶点亲和力模型,在药物分子图的特征提取过程中引入自注意机制,充分提取其有效特征表征。将二维分子图中每个原子的特征根据注意力得分进行加权,然后聚合为分子表示,并在基准数据集上实现集中式和分布式两种不同的池化架构进行比较。此外,在处理蛋白质序列的过程中,受GDGRU-DTA中蛋白质特征提取方法的启发,我们继续将蛋白质序列解释为时间序列,并利用双向长短期记忆(BiLSTM)网络提取其特征,因为长氨基酸序列具有上下文依赖性。同样,DSAGLSTM-DTA在蛋白质特征提取过程中也利用了自注意机制来获得蛋白质的综合表征,将批中元素的最终隐藏状态与LSTM的每个单元输出进行加权,并将结果表示为蛋白质的最终特征。最后,将药物和蛋白质的表示连接到预测块中进行最终预测。在不同的回归数据集和二值分类数据集上对该模型进行了评价,结果表明,DSAGLSTM-DTA模型优于一些现有的DTA模型,具有良好的泛化能力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
A Machine Learning Method for Prediction of Yogurt Quality and Consumers Preferencesusing Sensory Attributes and Image Processing Techniques Automatic Spectral Classification of Stars using Machine Learning: An Approach based on the use of Unbalanced Data Ai_Birder: Using Artificial Intelligence and Deep Learning to Create a Mobile Application that Automates Bird Classification DSAGLSTM-DTA: Prediction of Drug-Target Affinity using Dual Self-Attention and LSTM Multilingual Speech to Text using Deep Learning based on MFCC Features
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1