Table 5 Introduction to Deep Learning Methods
From: Application of Artificial Intelligence In Drug-target Interactions Prediction: A Review
Author | Year | Dataset | Method description | Characteristic |
---|---|---|---|---|
Zhao et al.78 | 2020 | DrugBank | Improved graph representation learning with multiple data. | The information related to drugs, targets and drug-target pairs are merged into the heterogeneous graph. This method enriches the meaning of node information and optimizes the excessive smoothing phenomenon in traditional graph neural networks. |
Monteiro et al.79 | 2020 | Private dataset | Based on Convolutional Neural Network (CNN) with inputs as protein sequences and structured compounds SMILES. | This approach is preferable to obtaining a table of data from a traditional descriptor. |
Zhao et al.50 | 2020 | Davis, KIBA | A semi-supervised approach based on Generative Adversarial Networks (GANs) for predicting binding affinities. | The method consists of two parts, two GANs for feature extraction and a regression network for prediction, which is the first method based on semi-supervised GANs to predict binding affinity. |
Jin et al.36 | 2021 | KIBA | Enhanced molecular expression prediction of drug-target interactions by sequence embedding and graph convolutional networks. | For protein sequences, feature embeddings of amino acids are pre-trained using language modeling and fed into a convolutional neural network model for representation learning. |
Wang et al.80 | 2021 | DUD-E, human, BindingDB | Multitasking neural network consisting of a graph neural network, an attention module, and a multilayer perceptron. | Good multitasking performance, excellent in both binding affinity prediction and interaction categorization. |
Huang et al.81 | 2021 | Biosnap, Davis, BindingDB | The method specifically selects a knowledge-heuristic pattern mining algorithm for the substructures and models the interaction of the data on this basis. An enhanced transformer encoder is also employed in unlabeled data to capture information about its substructures. | Fully exploiting parsing of substructures and potential knowledge of unlabeled data. |
Cheng et al.82 | 2021 | human, C.elegans, DUD-E, DrugBank | Based on graph neural networks, the key binding features of sequences are searched for through two engagements of the multiple attention mechanism. | Convolutional neural network is affected by the size of the convolutional window, and can only extract the local correlation information in the window area, but not the long-range contextual relationship. This method can solve the long-range dependency problem of sequences to some extent. |
Peng et al.44 | 2021 | Luo and Yamanishi | An end-to-end heterogeneous graph learning architecture. | The input to this heterogeneous network contains biological information in multiple modalities, and a low-dimensional feature representation of the data is learned end-to-end in model training. |
Wang et al.83 | 2022 | Private dataset | Construction of a knowledge graph of drug-target pairs. | Knowledge graph to get recommendation information. |
Li et al.45 | 2022 | Luo | Graph neural networks are utilized to synthesize node attributes and topological information. | The extracted low-dimensional features are more representative, taking into account the nodes’ own attributes and the relationships between nodes. |
Yu et al.47 | 2022 | Luo | A framework for predicting drug-target interactions by information aggregation based on heterogeneous graphical neural networks with incorporation of attention mechanisms. | The method first obtains the molecular fingerprint information of the drug and the pseudo-amino acid composition information of the protein, and then extracts the initial features of the nodes by Bi-LSTM and aggregates the heterogeneous neighbors using the attention mechanism. |
Wang et al.52 | 2022 | Luo | A method for predicting drug-target interactions based on the convolution of heterogeneous network graphs integrating network information on drug-target interactions, drug-drug interactions, drug-drug similarity, target-target interactions, and target-target similarity. | Graph convolution operations are performed in the heterogeneous network to obtain node embeddings of drugs and targets. An attention mechanism is also introduced between the graph convolution layers to jointly process the node embeddings from each layer. The model uses fewer network types and has high predictive performance. |
Nguyen et al.53 | 2022 | PDBbind, Fragalysis | A method based on graph neural networks that introduces a multi-hop gating mechanism, as well as an atomic enhancement design for data in the face of insufficiently rich coronavirus datasets. | Better atomic representations can be obtained from non-neighbor information to improve the training effect of the model, while the authors mentioned that if a large dataset can be introduced to do migration learning can improve the model performance to better. |
Zhao et al.84 | 2022 | Kinase, Davis, Metz, KIBA | Deep sequence learning for drug target binding affinity prediction based on attention mechanisms. | Leverage attention mechanisms to focus on important key subsequences in drug and protein sequences to predict affinity. |
Zhang et al.4 | 2022 | Tang, DrugBank, KEGG and PubChem | Transformer-based deep learning model with convolutional network and graph neural network can effectively extract the local residual information in the sequence, and the interaction data of molecular structure. | Â |
Wang et al.85 | 2022 | BindingDB | A multi-view strategy with converters is used to achieve dimensionality reduction for matrix dimensions, while an end-to-end model is used to reduce the complexity of the model with feature extraction steps. | Simplifies the characterization learning task for the acquisition of locally important residues. |
Huang et al.38 | 2022 | human, C.elegans, BindingDB | A Transformer-based shared attention mechanism model for extracting desired features from raw protein sequences and drug molecule SMILEs, respectively. | Apply pre-trained encoders to protein coding to solve the problem of scarce labeling data. And introduce migration learning as a pre-training guide. |
Zhao et al.40 | 2022 | DrugBank, Davis, KIBA, BindingDB | In the module used for local feature extraction of sequences, CNN and Transformer are combined to be able to capture and encode feature information over long distances, and then finally unite the global information with the local features as well as inter-class correlation information. | By dividing the drug, target and interaction information into three modules and extracting the key features separately, the local pattern information in the sequence can be better learned. |
Wang et al.86 | 2022 | human, C.elegans, Davis | Combines graph convolutional neural networks and transformer methods with a multi-head attention mechanism to focus on atomic structure information. | The ability to learn conformationally stable atomic information demonstrates its interpretability at the atomic level. |
Li et al.87 | 2022 | human, DUD-E, BindingDB | A bipartite interaction messaging module has been used to capture the bidirectional effects of drug-target interactions, and the weight display has been optimized for better interpretability. | The importance of focusing on both sides of the drug and the target. |
Ranjan et al.88 | 2022 | MOSES | Combining a graph neural network approach with a knowledge graph approach and using an early fusion approach to representational screening of protein tertiary structure and order predicts the binding affinity scores of the generated molecules. | This combined approach effectively reduces the number of generated molecules and retains most of the theoretically well-bound molecules. |
Liu et al.37 | 2022 | DUD-E, LIT-PCBA | An intermolecular map conversion method that incorporates a special attention mechanism used to find topological and spatial information between molecules, and uses a three-way transformer to model intermolecular information. | This approach excels at predicting binding positions and has been validated for drug screening of coronaviruses. |
Ye et al.89 | 2022 | the gold standard dataset, DrugBank | A graph autoencoder and multisubspace deep neural network. | Enhancing its learning capabilities through automatic coding of graphs, subspace layers, and integration layers allows for more features to be extracted from network inputs and for better training of DNN networks. |
Bae et al.90 | 2022 | Davis, BindingDB | A DTA prediction model that can take into account local substructures. It incorporates an attention mechanism between drugs and proteins that focuses on local information while focusing on global information. | Attention has been given to the interrelationships between protein and drug substructures, and it is thought that it is the interaction between certain specific substructures that is critical for the binding of these drugs and targets. |
Joshy et al.91 | 2022 | KIBA | Based on CNN, the inputs are the FASTA of the protein and the neural fingerprint of the drug, and a multilayer interaction network of the two is constructed to realize the prediction of DTA. | This approach effectively realizes the embedding of drug molecule structure by focusing on chemical properties such as drug molecule structure during drug data processing. |
Mehdi et al.92 | 2022 | DUD-E, Human, BindingDB | A graph deep learning model that incorporates an attention mechanism tries to understand this internal semantic-level association and external intermolecular association of drugs and targets from an NLP perspective. | Interpretation of drug-target associativity relationships as sentence-level relationships. Stronger interpretability in the context of applying NLP methods. |
Wu et al.93 | 2022 | Drugbank, KEGG | A knowledge graph attention network that decomposes the representation of relationships between nodes with a knowledge graph and adds an attention weighting scheme to filter out important features. | The method cleverly converts the DTIs classification prediction problem into a neighborhood node linkage prediction problem for knowledge graphs. |
Li et al.51 | 2022 | Hetero-A, Hetero-B | A heterogeneous graph attention network with an additional graph attention diffusion layer. | More indirect node information is captured in the intra-layer and inter-layer perspectives, thus expanding the perceptual field and realizing attentional perception enhancement. The method also found that the effect of expanding the perceptual field is different to different degrees in the intra-layer and inter-layer perspectives, and when the perceptual field is expanded to the same degree, the inter-layer perceptual field can be expanded to achieve better model prediction than the intra-layer perceptual field. |
Wang et al.94 | 2022 | Davis, Biosnap, DrugBank | A parallel decoupling method combining CNN and Transformer. | Local and global features are extracted separately and coupled by a cross-attention mechanism, while the interaction information of drug-target pairs is convolved with a dynamically generated filter to achieve a better representation for local and global features. |
Qu et al.46 | 2022 | DrugBank, HPRD | A new model for drug-target interaction prediction based on heterogeneous network graph embedding. | Learning nodes (drugs, proteins) and their topological neighborhood representations by extracting higher-order structural information using a GCN-inspired graph autoencoder to form heterogeneous networks. |
Ma et al.95 | 2022 | PubChem, DrugBank, UniProt | Adding multiple self-attention mechanisms to both drug processing and drug-target co-processing phases. | Multiple attention mechanisms can help improve the model’s handling of specific requirements. |
Yin et al.43 | 2023 | DrugBank, TwoSides dataset, DDInter | A generalized deep learning framework for graph-based prediction of drug interactions and drug target interactions. | Learning the structures and sequences of drugs and proteins using Residual Graph Convolutional Networks (RGCNs) and Convolutional Networks (CNNs) to improve the prediction accuracy of Drug-Drug Interaction (DDI) and DTI. |
Zhang et al.54 | 2023 | DrugBank, UniProt, MalaCar | Two different graph autoencoders are used to form an enhanced deep autoencoder. | Feature inference and label propagation between known DTIs for spaces containing unknown drug targets. |
Bian et al.96 | 2023 | Davis, KIBA, DrugBank | A multiple cross-attention mechanism based on weight sharing is employed, using only sequence information as input. | The 2D feature maps of proteins and drugs are compressed into 1D feature maps, which are then connected and fed into FCN for classification. It can effectively solve the class imbalance problem and overfitting problem in drug-target datasets. |
Wen et al.97 | 2023 | Human, C.elegans, GPCR, Davis | Extraction of long-range interdependent features of sequences using a multi-attention mechanism to extract atom-amino acid interaction features. | Focus on the complex interactions between internal atoms and macromolecular compounds. |
Kavipriya et al.98 | 2023 | KEGG BRITE, BRENDA, SuperTarget, DrugBank | An optimal recurrent neural network-based approach using a semi-supervised approach based on BiLSTM model of RNN to handle inter-drug and inter-target interactions and initiated as weights. | Since the performance of the BiLSTM technique is strongly influenced by the hyperparameters, the Adam optimizer is used to adjust the hyperparameters during the model training process to improve the prediction performance. |
Zheng et al.39 | 2023 | DrugBank, BindingDB | A multi-task learning mechanism with two transformer encoder-decoder structures is introduced to enhance the extraction of target features. | The first to apply Mol2Vec, BERT, attentional mechanisms and multitasking mechanisms to a single model. |
Zhu et al.99 | 2023 | Davis, KIBA | Sequence features and substructural features of individual molecules are extracted by the self-coder to express relationships, while information interaction paths are established between molecules and sequences, forming an associative learning mechanism. | The correlation between substructures is better expressed, which improves the accuracy of the predicted values for the DTA task. |
Sofia et al.100 | 2023 | Davis, KIBA | Based on CNN, inputs are in SMILES and protein residue information. | There is no need to construct a similarity matrix, thus improving computational efficiency compared to some machine learning models. |
Chen et al.56 | 2023 | DrugBank, HPRD, CTD, SIDER | A graph representation learning method that embeds nearest neighbor relationships between nodes to keep the topology of the relationships between nodes undeformed in the graphical learning process while maintaining the graph structure. | Effectively solves the problem of large number of isolated nodes in traditional graph convolution autoencoder models. |