Next Article in Journal
MIMR: Development of a Web-Based System for miRNA and mRNA Integrated Analysis
Previous Article in Journal
Non-HLA Autoantibodies Against Angiotensin II Receptor 1 (AT1R) and Endothelin A Receptor (ETAR) in Pediatric Kidney Transplantation
Previous Article in Special Issue
Prediction of Drug-Target Affinity Using Attention Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

NFSA-DTI: A Novel Drug–Target Interaction Prediction Model Using Neural Fingerprint and Self-Attention Mechanism

School of Information, Guizhou University of Finance and Economics, Guiyang 550025, China
*
Authors to whom correspondence should be addressed.
Int. J. Mol. Sci. 2024, 25(21), 11818; https://doi.org/10.3390/ijms252111818
Submission received: 27 September 2024 / Revised: 24 October 2024 / Accepted: 1 November 2024 / Published: 3 November 2024

Abstract

:
Existing deep learning methods have shown outstanding performance in predicting drug–target interactions. However, they still have limitations: (1) the over-reliance on locally extracted features by some single encoders, with insufficient consideration of global features, and (2) the inadequate modeling and learning of local crucial interaction sites in drug–target interaction pairs. In this study, we propose a novel drug–target interaction prediction model called the Neural Fingerprint and Self-Attention Mechanism (NFSA-DTI), which effectively integrates the local information of drug molecules and target sequences with their respective global features. The neural fingerprint method is used in this model to extract global features of drug molecules, while the self-attention mechanism is utilized to enhance CNN’s capability in capturing the long-distance dependencies between the subsequences in the target amino acid sequence. In the feature fusion module, we improve the bilinear attention network by incorporating attention pooling, which enhances the model’s ability to learn local crucial interaction sites in the drug–target pair. The experimental results on three benchmark datasets demonstrated that NFSA-DTI outperformed all baseline models in predictive performance. Furthermore, case studies illustrated that our model could provide valuable insights for drug discovery. Moreover, our model offers molecular-level interpretations.

1. Introduction

Despite significant advances in basic life sciences and biotechnology, the drug discovery and development process is still limited by slow time and high costs [1]. The average duration for creating a small-molecule medication is approximately 15 years, costing around USD 2 billion [2]. While clinical studies are widely acknowledged as a crucial component of drug development, the greatest potential for time and cost savings lies in the earlier discovery stage [3,4]. The prediction of drug–target interaction (DTI) plays a pivotal role in guiding research and development efforts, making it an essential part of the drug discovery process [5,6].
Recently, significant advancements in drug–target interaction prediction have been achieved due to rapid progress in deep learning (DL)-based approaches [7,8,9,10]. Many of these approaches seamlessly integrate the drug chemical space, protein sequence, and interaction information into a comprehensive end-to-end framework [11,12]. They approach DTI prediction as a binary classification task and utilize various deep encoding and decoding modules, such as graph neural networks [13,14,15], deep neural networks [16,17], or transformer architectures [18,19], to make predictions. With the advancements in deep learning techniques, these models can automatically learn data-driven representations of drugs and proteins from extensive DTI data instead of solely relying on predefined descriptors.
However, the current DL-based models, despite exhibiting decent performance, are subject to two limitations. One limitation arises from the predominant focus on local feature extraction, neglecting the acquisition of global representation [13,14,16]. The prediction of drug–target interaction necessitates comprehensive consideration of various aspects pertaining to receptor and ligand information [20]. Therefore, focusing only on local features tends to limit the modeling power and forecasting performance of the model. The second limitation arises from the inability to explicitly model and learn the local key interaction sites between the drug molecule and target protein. This limitation impedes the ability to provide a more intuitive and molecular-level interpretability of the mechanisms underlying drug–target interaction.
In recent years, the combination of molecular fingerprinting and deep learning models has further improved the performance of models in dealing with complex molecular structures [21]. For example, Zhu et al. [17] proposed a fingerprint-embedding framework for drug–target binding affinity prediction (FingerDTA), which uses convolutional neural networks (CNNs) to extract local patterns and utilizes fixed fingerprints to characterize global information. Lee et al. [16] proposed a novel model called DeepConv-DTI, which captures the local residue patterns by convolving amino acid subsequences of various lengths, and it uses a fully connected neural network to encode the fixed ECFP4 [22] drug fingerprint. On the one hand, these traditional fingerprints are usually generated based on the local structure of the molecule (such as specific atoms or chemical substructures) and rely on predefined rules, so they cannot effectively capture the overall topology of the molecule [23]. In contrast, the neural fingerprint [24] based on a graph neural network can gradually integrate local information into global information by recursively aggregating the drug features of all nodes. On the other hand, the limitation of CNN as the protein encoder is that the dependencies between distant amino acids in the protein sequence cannot be captured well [25]. Fortunately, the self-attention mechanism can assign different weights to each position in the sequence, allowing it to capture dependencies between positions more effectively [26]. This offers the potential to enhance CNN’s capability in capturing the global information within the sequence.
Building on the advantages of neural fingerprints, Duvenaud et al. [24] integrated them into graph neural networks (GNNs) to propose the neural fingerprint graph neural network (NFGNN). Originally designed for predicting molecular properties, NFGNN has been shown through comparative experiments by Feldmann et al. [27] to enhance the generalization ability of molecular property predictions, demonstrating robust performance even for molecules that differ from the training set. Recently, the application of NFGNN in DTI prediction has begun to make progress. For instance, Joshy et al. [28] employed NFGNN in the search for drugs to treat orphan diseases, validating its effectiveness. However, the application of NFGNN in DTI prediction is quite scarce. The full potential of NFGNN in this area remains to be further explored.
In recent years, the self-attention mechanism has increasingly been incorporated into prediction models in the DTI field due to its advantages [29]. For instance, compared to DrugBAN developed by Bai et al. [14], CAT-DTI proposed by Zeng et al. [15] has introduced further innovations in the protein encoder module. They integrate the transformer architecture with CNN, enhancing the feature extraction capabilities of CNNs through the self-attention mechanism. Although CAT-DTI demonstrates superior performance compared to existing baseline models, the combination of these two architectures can complicate the training process and hyperparameter tuning. Furthermore, CAT-DTI does not adequately capture local site-specific interactions in drug–target pairs, which limits its interpretability. Therefore, further research is needed to explore the scientific integration of self-attention mechanisms into prediction models and to improve model interpretability.
Therefore, in order to address the aforementioned limitations and explore potential areas for improvement, we propose a drug–target interaction prediction model (NFSA-DTI) based on neural fingerprint and a self-attention mechanism. In conclusion, the key contributions of our work can be summarized as follows:
We propose an enhancing self-attention convolutional module (ESACM) that utilizes the self-attention mechanism to enhance CNN’s ability in capturing the long-distance dependencies between the subsequences in the target amino acid sequence. ESACM enables the protein encoder to comprehensively calculate the impact of subsequences at different positions on the target sequence.
We have successfully applied the neural fingerprint graph neural network (NFGNN) to the drug–target interaction prediction task and verified its effectiveness. In contrast to fixed fingerprints, which require extensive lookup tables to uniquely encode all possible molecular structures, neural fingerprints can encode all molecular structures using trainable parameters.
We propose the integration of an attention pooling method into the bilinear attention network and demonstrate its effectiveness. This pooling method can assign weights to each element in the input matrix based on its importance, thereby augmenting the model’s learning ability of the local pivotal binding sites within the drug–target pair.

2. Results and Discussion

2.1. Model Performance Comparison

We compared NFSA-DTI with seven baseline models in three benchmark datasets. The results of the comparative experiments are presented in Table 1.
In terms of AUROC, AUPRC, and accuracy, NFSA-DTI was better than the baseline models according to the results in Table 1. In the BindingDB dataset, compared with the currently best-performing CAT-DTI, AUROC, AUPRC, and accuracy of NFSA-DTI were improved by 0.5, 1.0, and 1.1 percentage points, respectively. We believe that the stronger global feature extraction capability, as well as the introduction of multiple attention mechanisms, motivated NFSA-DTI to perform better than CAT-DTI. However, the performance gap between NFSA-DTI and CAT-DTI in the BioSNAP dataset was not particularly obvious. This phenomenon may be caused by the fact that on a balanced and data-heavy BioSNAP dataset the model may be more prone to overfitting during training because it must simultaneously process a large number of samples from various classes, and the features between different classes may interfere with each other, thus affecting the overall performance. To address this issue, we modified the number of neurons per layer in both the NFGNN and CNN models, as well as adjusted their dropout rates, to reduce the risk of potential overfitting. These modifications are intended to further narrow the performance gap with the CAT-DTI model in terms of AUROC and AUPRC.
In terms of sensitivity and specificity, NFSA-DTI did not show complete advantages over the baseline models. This may be because the model tended to predict a certain class of samples, thus inevitably sacrificing a part of the sensitivity to the minority class or the specificity to the majority class. However, the model performed well on AUROC and AUPRC, which indicated that the model could still effectively distinguish between positive and negative samples in general. Overall, NFSA-DTI outperformed all baseline models under most evaluation metrics, thus demonstrating the effectiveness of our proposed model.

2.2. Ablation Study

In order to verify the influence of each innovation module in NFSA-DTI on the model’s performance, we designed several model variants to conduct ablation experiments. The results are shown in Figure 1.
The three ablation models are explained as follows:
(1) w/o ESACM: Protein encoder exclusively employs a three-layer CNN without the addition of the self-attention enhancing unit (others same as NFSA-DTI).
(2) w/o NFGNN: Drug encoder employs a three-layer GCN instead of a three-layer NFGNN (others same as NFSA-DTI).
(3) w/o Attention pooling: Sum pooling is used for the final pooling of the bilinear attention network instead of attention pooling (others same as NFSA-DTI).
According to the results in Figure 1, compared with w/o ESACM, NFSA-DTI achieved a 1.18 percentage point increase in AUROC and a 1.68 percentage point increase in AUPRC on the Human dataset, and a 1.73 percentage point increase in AUROC and a 1.92 percentage point increase in AUPRC on the BioSNAP dataset. This is because, compared with the pure use of CNNs as the protein encoder, ESACM could use the self-attention mechanism to effectively enhance the ability of CNNs to capture the long-distance dependencies between the subsequences in the target amino acid sequence, thus further improving the performance. Compared with w/o NFGNN and w/o Attention pooling, NFSA-DTI also had relatively obvious advantages, which also indicated that compared with using GNN alone as the drug encoder, NFGNN used the neural fingerprint encoding the entire molecular structure as an additional global information supplement. In addition, compared with the fixed fingerprint, the neural fingerprint could be obtained by parameter learning and had more efficient retrieval advantages. Compared with sum pooling as the final pooling operation of the bilinear attention network, attention pooling could enhance the model’s ability to learn the local important binding sites in a drug–target pair.

2.3. Case Study

In order to verify our model’s ability to discover new drugs and new target proteins, we constructed a new test dataset from the DrugBank online database, consisting of drug–target pairs absent in the BindingDB dataset, and we used the model trained on the BindingDB dataset to conduct the test. Taking Vascular Endothelial Growth Factor Receptor 2 and Nortriptyline as examples, according to the interaction probability score p, the predicted top 10 drug and target protein candidates are shown in Table 2 and Table 3, respectively.
As can be seen in Table 2 and Table 3, our model successfully predicted three drugs and two proteins (marked in bold) that could be validated in the DrugBank online database and two drugs and four proteins (marked with underline) that could be validated in the TTD [36] online database. The pharmacological effects of other drug–target pairs in the above two tables were currently unknown in the DrugBank and TTD databases and need to be further verified. It is worth mentioning that the drug candidate or target protein candidate with the highest prediction score in the table was consistent with the ranking recorded in the DrugBank database.
Based on the unknown drug–target interaction pairs mentioned above, researchers can discover new drug candidates that may have therapeutic potential, especially for those diseases that have not been fully studied [37]. For example, in Table 2, 2-aminobenzimidazole would potentially act on Vascular Endothelial Growth Factor Receptor 2 as a new drug. Moreover, for already approved drugs, this will help to search for their possible new mechanisms of action and new therapeutic applications, thereby helping researchers to explore the potential of these drugs in the treatment of other diseases [38]. Similarly, in Table 3, Nortriptyline may act as a repositioning drug on a new target Alpha-2A adrenergic receptor. Although these unknown predictions require further experimental validation, they provide new directions and possibilities for drug research.

2.4. Interpretability Analysis

Another advantage of NFSA-DTI is that it provides molecular-level interpretations essential for drug design efforts. For example, in the Human test dataset, we randomly selected a drug–target interaction pair and plotted the 2D visualization result of its drug molecule as shown in Figure 2a, and we further verified our result by using the Protein Data Bank (PDB) [39] online database, as shown in Figure 2b,c.
According to Figure 2a, our model accurately predicted the involvement of the carboxylate group of mycophenolic acid in receptor binding. This prediction was further validated in Figure 2b, where both oxygen atoms of the carboxylate group were shown to act as hydrogen bond acceptors. One oxygen atom interacts with the back-bone of Ser275 and Ser276, while the other forms a specific interaction with the sidechain of Ser276. Additionally, in Figure 2a, our model predicted that the hydroxyl and methyl groups on the phenyl ring, as well as the carbonyl oxygen atom, are involved in receptor binding. This prediction is confirmed in Figure 2b, where the hydroxyl and methyl groups act as hydrogen bond donors, forming interactions with the sidechains of Thr333 and Asp274, respectively. The carbonyl oxygen atom functions as a hydrogen bond acceptor, interacting with the sidechains of Thr333 and Cys331. Unfortunately, the oxygen atom of the epoxy group was mistakenly predicted to be involved in the receptor interaction. Nonetheless, the overall interpretability of our model provides significant reference value.
In Figure 2b, the dark blue ligand exposure area is mostly concentrated near the carboxylate group of mycophenolic acid. In addition, the light blue receptor exposure area is mainly concentrated near Ser276. Combined with the orange area in Figure 2a indicating possible interaction, it can be inferred that the carboxylate group’s oxygen negative ion and the oxygen atom of the carboxylate group are the most important in this interaction. The 3D visualization of Figure 2c, as a supplement to the 2D visualization of Figure 2b, will help researchers analyze the 3D structure of mycophenolic acid and inosine monophosphate dehydrogenase acceptor and identify key pharmacophore and binding site.

3. Materials and Methods

3.1. Datasets

We evaluated NFSA-DTI using three publicly available DTI datasets: BindingDB, BioSNAP, and Human. Information about them is shown below.
The BindingDB Dataset [41] is a large drug–target dataset, containing thousands of small molecule drugs and protein targets. The targets also cover different species, but mainly focus on human targets. It is worth mentioning that BindingDB is unbalanced in terms of dataset distribution.
The BioSNAP dataset, developed by Huang et al. [18] and Zitnik et al. [42] from the DrugBank database [43], is a balanced dataset. It includes verified positive samples as well as an equal number of randomly paired negative samples that have never been encountered before. This dataset considers interactions between small chemical drugs and target proteins, all of which have been experimentally validated via biological experiments or formal pharmacological studies.
The Human dataset, constructed by Liu et al. [44], is a balanced dataset, incorporating high-confidence negative samples obtained through silicon screening methods.
Table 4 gives the detailed statistics of these three datasets. In the table, Interactions* are defined as the total number of interacting sample pairs and non-interacting sample pairs in the dataset. P2N represents the ratio of the number of interacting sample pairs to non-interacting sample pairs in the dataset.
The positive samples in the three datasets are based on the literature and experimental verification. However, due to the relative scarcity of experimentally verified negative sample data, researchers usually need to adopt different strategies to construct negative sample sets [9,16]. For the Human dataset, based on the existing biological data and chemical characteristics, the researchers used silicon screening computing technology to efficiently screen out the combinations with low correlation as negative samples, so as to improve the confidence of negative samples. Regarding the BioSNAP dataset, using the biological network construction method, the researchers identified pairs of nodes that were not directly connected in the network and randomly selected negative samples from them to ensure the confidence of the negative samples. For the BindingDB dataset, researchers generated negative samples by excluding known drug–target binding pairs and rigorously screened negative samples to ensure that these combinations did not overlap with positive samples.

3.2. Baselines

In order to comprehensively evaluate the prediction performance of our model, we chose seven representative baseline models, including two machine learning models (SVM [45] and RF [46]), and five deep learning models (DeepConv-DTI [16], GraphDTA [13], MolTrans [18], DrugBAN [14], and CAT-DTI [15]). Information about them is shown below.
SVM: By learning the optimal hyperplane, the interaction between drugs and targets can be effectively distinguished in the high-dimensional feature space, which has a relatively strong classification ability and good generalization performance.
RF: By integrating multiple decision trees, the interaction between drugs and targets is predicted in a voting manner, which has strong anti-noise and robustness. It performs well when dealing with high-dimensional data, but may be vulnerable to uneven feature importance.
DeepConv-DTI: By using a convolutional neural network, amino acid subsequences of various lengths are convolved to capture local residue patterns, and a fully connected neural network is used to encode the fixed ECFP4 drug fingerprint. It outperforms previous models based on protein descriptors.
GraphDTA: By representing drugs as graphs and using the graph neural network to predict the affinity of the drug to the target, it can effectively process the topological structure data of drug molecules and improve the prediction accuracy.
MolTrans: By introducing the self-attention mechanism of the transformer, the drug molecules and protein sequences are embedded into a unified vector space, in order to effectively capture the interaction characteristics between them. It has high flexibility and performance in dealing with complex molecular relationships.
DrugBAN: By introducing a bilinear attention network, the interaction strength between the drug and the substructure of the target will be embedded into the bilinear attention matrix for downstream prediction tasks. It can better capture the local feature correlation and improve the performance of the model.
CAT-DTI: By combining the graph convolutional neural network, a transformer architecture, and the cross-attention mechanism, it can effectively capture the information of drug and target sequences and improve the prediction performance. It has an advantage when dealing with long target sequences and is better able to model complex interactions.

3.3. Metrics

In our experiment, the AUROC (area under the receiver operating characteristic curve) and AUPRC (area under the precision-recall curve) serve as primary metrics for assessing model classification performance. AUROC measures the model’s ability to discriminate between positive and negative samples across different classification thresholds. It considers both the true positive rate (sensitivity) and the false positive rate (1—specificity). A higher AUROC value indicates better discriminative power and overall performance. AUPRC, on the other hand, focuses on the trade-off between precision and recall. It measures the model’s ability to correctly identify positive samples while minimizing false positives. In addition, we adopted accuracy, sensitivity, and specificity under the optimal F1 score threshold as additional evaluation metrics.

3.4. Implement Details

To optimize the implementation of NFSA-DTI, we randomly split each dataset into train/validate/test sets at a ratio of 0.7/0.1/0.2. The test set is employed to evaluate the model’s final performance and is entirely independent of the training and validation processes, thereby ensuring the objectivity and accuracy of the evaluation results. Compared to the 7:2:1 division, the 7:1:2 division enhances the reliability of the model’s final evaluation. We used the Human dataset to train the NFSA-DTI model and determined the initial combination of hyperparameters. Subsequently, we fine-tuned the parameters and trained NFSA-DTI. When NFSA-DTI exhibited superior performance on the validation set of each dataset, we determined the optimal values of the hyperparameters. The partial learning curves for the BindingDB validation set are depicted in Figure 3, illustrating the influence of varying certain hyperparameter values. The figure demonstrates that the optimal performance of the model is achieved when the learning rate of the optimizer reaches 5 × 10−5 and the number of attention heads in the bilinear attention network is set to 2. The main hyperparameters configuration is shown in Table 5. In this study, we conducted our experiments using Python 3.8 as the programming language. The deep learning framework used is PyTorch version 1.12, and we trained the model using an RTX 4090 24 G, accelerating the computation using NVIDIA CUDA version 11.6.

3.5. Methods

Figure 4A shows the framework of our proposed NFSA-DTI model.

3.5.1. Problem Formulation and Prerequisites

The objective of DTI prediction is to develop a model M, which effectively maps the combined feature representation space P × G of a protein sequence P and a drug molecular graph G, yielding an interaction probability score p [ 0 , 1 ] .
The target protein is represented by one amino acid sequence denoted as P = ( a 1 , , a n ) , where each token corresponds to one of the 23 amino acids. The maximum length of P is set to 1200, which is long enough to cover common amino acid sequences while remaining computationally and storage efficient. Amino acid sequences that exceed this threshold are truncated. For amino acid sequences that are not of length 1200, zeros will be padded at the end of the sequence. The amino acid sequences [47] are a standardized representation of proteins that can be expressed and processed in a computable way.
Major current deep learning methods for predicting drug–target interaction typically use the simplified molecular input line entry system (SMILES) [48] as an input representation. The limitation of SMILES lies in its one-dimensional nature, which hinders its ability to capture the intricate spatial arrangement of molecules, potentially resulting in the loss of crucial chemical information. However, in our model, the input SMILES are converted into the corresponding two-dimensional graph structure data G. It creates actual nodes and edges based on the drug molecule’s atoms and chemical bonds, with virtual nodes and self-loop edges filling in the rest. Node count is capped at 290.

3.5.2. Protein Feature Encoder

The conventional CNN architecture exhibits certain limitations in handling long-distance dependencies of sequences due to the restricted receptive field of convolution operations [49]. In contrast, the self-attention mechanism enables each position to interact with all other positions within the tensor, granting access to comprehensive information about the entire tensor [50]. Therefore, we propose ESACM to augment model performance. ESACM integrates a self-attention mechanism that enhances the CNNs’ ability to capture the long-distance dependencies between the subsequences in the target amino acid sequence. The framework diagram of ESACM is shown in Figure 4B.
Building upon the concept of word embedding for amino acid sequences by Bai et al. [14], the amino acids are initialized into a learnable embedding matrix E p R 23 × D p , where 23 is the number of amino acid types and D p is the latent space dimension. By looking up E p , each protein sequence P can be initialized to the corresponding feature matrix X p R Θ p × D p , where the parameter Θ p represents the maximum permissible length of a protein sequence, which is generally 1200 in our experiment, and each row of the matrix denotes a residue representation in the target amino acid sequence. In the initial convolution layer, the feature matrix X p will be input into the first CNN for one-dimensional convolution to extract amino acid sequence features, and this process utilizes a convolution kernel sized 3 × 1 and a stride of 1, such as “GSHMAS…PQQG” ⟶ “GSH”, “SHM”, “HMA”, “MAS”, …, “PQQ”, and “QQG”. The next two layers keep the same stride as the first, but further enlarge the kernel size to acquire more features of local subsequences. The convolutional operations can be expressed by the following formula:
H p ( l + 1 ) = σ ( CNN ( W c ( l ) , b c ( l ) , H p ( l ) ) ) ,
where H p ( l + 1 ) is the l-th hidden protein representation and H p ( 0 ) = X p . σ ( · ) denotes ReLU ( · ) . W c ( l ) and b c ( l ) are learnable weight matrices (filters) and bias vector in the l-th CNN layer.
The output of the third convolutional layer will subsequently undergo computation in the self-attention enhancing unit. Specifically, the self-attention enhancing unit consists of two stages, corresponding to the following formulas.
Stage I:
q i , j ( l ) = W q ( l ) f i , j , k i , j ( l ) = W k ( l ) f i , j , v i , j ( l ) = W v ( l ) f i , j .
Stage II:
p i , j = l = 1 N a , b N m ( i ) A ( q i , j ( l ) , k a , b ( l ) ) v a , b ( l ) .
The specific formula is presented as follows:
p i , j = l = 1 N a , b N m ( i ) A ( W q ( l ) f i , j , W k ( l ) f a , b ) W v ( l ) f a , b ,
where p i , j is the value in row i and column j of the protein representation. ∏ is the concatenation of the outputs of N attention heads. N m ( i ) represents one column region whose length is the maximum column length m of the feature matrix, and its abscissa is i. A ( W q ( l ) f i , j , W k ( l ) f a , b ) is the corresponding attention weight with regard to the features within N m ( i ) . W q ( l ) , W k ( l ) , and W v ( l ) are the projection matrices for queries, keys, and values, respectively. f i , j H p ( 3 ) and H p ( 3 ) denote the output matrix of the third CNN. The attention weight is calculated as follows:
A ( W q ( l ) f i , j , W k ( l ) f a , b ) = softmax ( W q ( l ) f i , j ) ( W k ( l ) f a , b ) d ,
where d is the feature dimension of W q ( l ) f i , j .

3.5.3. Drug Feature Encoder

Graph neural network (GCN) [51] primarily updates node representations through local neighborhood information propagation, its ability to model the global structure is relatively limited. NFGNN, a type of neural network proposed by Duvenaud et al. [24], was originally designed for predicting molecular properties and has been proven effective in learning the representation of molecular graphs. Compared with GCN, NFGNN uses the neural fingerprint as an additional global information supplement. Moreover, in contrast to fixed fingerprints that necessitate a substantial number of lookup tables for encoding all possible molecular structures uniquely, neural fingerprints of NFGNN can encode the entirety of molecular structures using trainable parameters. Given the advantages of NFGNN, we therefore introduce NFGNN as the drug encoder of NFSA-DTI.
The flowchart of NFGNN is illustrated in Figure 4C. Specifically, the construction process of the neural fingerprint is represented by the following formulas:
r a = g ( a ) ,
v i = r i + j = 1 N r j , j N ( i ) ,
f = i = 1 M ( softmax ( σ ( v i H h ) W o ) ) ,
where a represents the node. g ( · ) is the encoding function of node features. v i represents the sum of features about the i-th node and its neighbors. N ( i ) is the set of neighbor nodes of the i-th node. H h is the hidden weight, and W o is the output weight. N denotes the number of neighbor nodes of the i-th node, and M denotes the total number of nodes in the graph. σ represents ReLU ( · ) . The subsequent message passing process can be expressed by the following formulas:
m i j ( l ) = ρ ( l ) ( h i ( l ) , h j ( l ) , h e i j ( l ) ) , j N ( i ) ,
m i ( l ) = ζ ( l ) ( m i j ( l ) | j N ( i ) ) ,
h i ( l + 1 ) = ϕ ( l ) ( h i ( l ) , m i ( l ) , f ) ,
where ρ ( · ) , ζ ( · ) , and ϕ ( · ) are functions of message construction, message aggregation, and vertex update of the l-th layer NFGNN, respectively. h i ( l ) is the node feature of the i-th node at the l-th layer, h j ( l ) is its neighbor node feature, h e i j ( l ) is the corresponding edge feature, and m is the message.

3.5.4. Bilinear Attention Network

We employ the bilinear attention network proposed by Bai et al. [14] and improve its subsequent pooling operation. Specifically, we replace the summation pooling with an attention pooling. The sum pooling method is relatively straightforward to implement and incurs a low computational cost. However, it lacks the ability to retain the intensity information of features, potentially compromising learning performance. To address this limitation and enhance the model’s capacity in capturing important information between drug substructures and target substructures, we employ an attention pooling approach inspired by Er et al.’s [52] investigation on the importance of sentence components. The attention pooling assigns weights based on the significance of each element in the input matrix, generating an enhanced pooled representation. The enhanced pooled representation promotes the model’s learning of the drug–target interaction modeling process, thereby improving the overall interpretability of the model. The overall flowchart of the improved bilinear attention network is shown in Figure 5.
Given the protein representation and drug representation H p ( 3 ) = { h p 1 , h p 2 , , h p M } , among them, h p 1 = { p 1 , 1 , p 1 , 2 , , p 1 , e } , , h p M = { p M , 1 , p M , 2 , , p M , e } , and H d ( 3 ) = { h 1 ( 3 ) , h 2 ( 3 ) , , h N ( 3 ) } via the protein and drug encoders, where M represents the maximum number of target protein substructures encoded by ESACM, N represents the maximum number of nodes in the drug molecular graph, and e represents the embedding dimension of the CNN. The bilinear interaction matrix I R N × M can be obtained as follows:
I = ( ( 1 · q ) σ ( ( H d ( 3 ) ) U ) ) · σ ( V H p ( 3 ) ) ,
where U and V are learnable weight matrices for drug and protein representations, q is a learnable weight vector, 1 is a fixed all-ones vector, and the symbols ∘ and σ represent the Hadamard (element-wise) product and ReLU ( · ) . The elements in I represent the intensity of interaction between drug–target substructures, indicating their potential binding sites. In order to intuitively understand bilinear interaction, element I i , j can be written as follows:
I i , j = q ( σ ( U h d i ) σ ( V h p j ) ) ,
where h d i is the i-th row of H d ( 3 ) and h p j is the j-th row of H p ( 3 ) , respectively, denoting the i-th and j-th substructural representations of the drug and protein.
In order to obtain the joint representation f , the interaction matrix I is processed by a bilinear pooling layer. Specifically, f is computed as follows:
f = σ ( ( H d ( 3 ) ) U ) · I · σ ( ( H p ( 3 ) ) V ) ,
where the weight matrices U and V are shared with the preceding interaction matrix layer to reduce the number of parameters and mitigate overfitting.
Subsequently, the final joint representation f will be obtained after the attention pooling, which is represented by the following formulas:
W i , j = Softmax ( b a tanh ( U a I i , j ) ) ,
f = i = 1 N j = 1 M ( W i , j · b a ) f ,
where b a and b a represent the bias vectors, U a denotes the weight matrix, and W i , j is the attention weight vector.

3.5.5. Fully Connected Classification

The interaction probability is computed by feeding the joint representation f into the decoder, which comprises a single fully connected classification layer followed by a sigmoid function, as follows:
p = Sigmoid ( W o f + b o ) ,
where W o is the learnable weight matrix and b o is the bias vector.

3.5.6. Backpropagation

The backpropagation process aims to minimize the cross-entropy loss [53] in order to optimize the model’s performance. The formula is presented as follows:
L = i ( y i log ( p i ) + ( 1 y i ) log ( 1 p i ) ) + λ 2 Θ 2 2 ,
where y i denotes the ground truth label for the i-th drug–target pair, p i represents its output probability according to the model, Θ represents all the weight matrices and bias vectors that can be learned, and λ is a hyperparameter used for L2 [54] regularization.

4. Conclusions

In this article, we propose a novel DTI prediction model called NFSA-DTI. In the protein encoder, ESACM captures the long-distance dependencies between the subsequences in the target amino acid sequence while extracting the local features of the sequence. In the drug encoder, NFGNN extracts features of the drug molecular graph via a message-passing mechanism, supplemented by the neural fingerprint as global information. In the feature fusion module, the bilinear attention network generates an enhanced pooled representation via attention pooling, improving the model’s ability to learn key local binding sites in the drug–target pair. The experimental results on three benchmark datasets demonstrated that NFSA-DTI outperformed all baseline models in terms of prediction performance. Further case experiments showed that the model has provided valuable insights for drug discovery efforts. Furthermore, the model provides more intuitive interpretability at the molecular level. In the future, we plan to extend our study by using other benchmark datasets.

Author Contributions

Conceptualization, Z.W., P.C. and F.L.; methodology, F.L., H.X., P.C. and Z.W.; software, F.L. and H.W.; validation, F.L., H.X. and S.L.; formal analysis, P.C. and Z.W.; investigation, F.L. and H.X.; resources, Z.W., P.C. and F.L.; data curation, F.L. and H.X.; writing—original draft preparation, F.L.; writing—review and editing, F.L., P.C. and Z.W.; visualization, F.L. and H.W.; supervision, P.C. and Z.W.; project administration, P.C. and Z.W.; funding acquisition, P.C. and Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 22103018 and No. 21963004) and the Top Talent Project of Guizhou Provincial Department of Education (QJJ-[2022]-080).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

NFSA-DTI and datasets used in this study are available on GitHub at https://github.com/xiaofeinb233/NFSA_DTI (accessed on 20 October 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gholap, A.D.; Uddin, M.J.; Faiyazuddin, M.; Omri, A.; Gowri, S.; Khalid, M. Advances in Artificial Intelligence in Drug Delivery and Development: A Comprehensive Review. Comput. Biol. Med. 2024, 178, 108702. [Google Scholar] [CrossRef] [PubMed]
  2. Sadybekov, A.V.; Katritch, V. Computational approaches streamlining drug discovery. Nature 2023, 616, 673–685. [Google Scholar] [CrossRef] [PubMed]
  3. Sarkar, C.; Das, B.; Rawat, V.S.; Wahlang, J.B.; Nongpiur, A.; Tiewsoh, I.; Lyngdoh, N.M.; Das, D.; Bidarolli, M.; Sony, H.T. Artificial intelligence and machine learning technology driven modern drug discovery and development. Int. J. Mol. Sci. 2023, 24, 2026. [Google Scholar] [CrossRef] [PubMed]
  4. Vora, L.K.; Gholap, A.D.; Jetha, K.; Thakur, R.R.S.; Solanki, H.K.; Chavda, V.P. Artificial intelligence in pharmaceutical technology and drug delivery design. Pharmaceutics 2023, 15, 1916. [Google Scholar] [CrossRef] [PubMed]
  5. Blanco-Gonzalez, A.; Cabezon, A.; Seco-Gonzalez, A.; Conde-Torres, D.; Antelo-Riveiro, P.; Pineiro, A.; Garcia-Fandino, R. The role of ai in drug discovery: Challenges, opportunities, and strategies. Pharmaceuticals 2023, 16, 891. [Google Scholar] [CrossRef]
  6. Singh, S.; Kaur, N.; Gehlot, A. Application of artificial intelligence in drug design: A review. Comput. Biol. Med. 2024, 179, 108810. [Google Scholar] [CrossRef]
  7. Askr, H.; Elgeldawi, E.; Aboul Ella, H.; Elshaier, Y.A.; Gomaa, M.M.; Hassanien, A.E. Deep learning in drug discovery: An integrative review and future challenges. Artif. Intell. Rev. 2023, 56, 5975–6037. [Google Scholar] [CrossRef]
  8. Tang, X.; Lei, X.; Zhang, Y. Prediction of Drug-Target Affinity Using Attention Neural Network. Int. J. Mol. Sci. 2024, 25, 5126. [Google Scholar] [CrossRef]
  9. Huang, Y.; Huang, H.Y.; Chen, Y.; Lin, Y.C.D.; Yao, L.; Lin, T.; Leng, J.; Chang, Y.; Zhang, Y.; Zhu, Z.; et al. A robust drug–target interaction prediction framework with capsule network and transfer learning. Int. J. Mol. Sci. 2023, 24, 14061. [Google Scholar] [CrossRef]
  10. Wang, X.; Liu, J.; Zhang, C.; Wang, S. SSGraphCPI: A novel model for predicting compound-protein interactions based on deep learning. Int. J. Mol. Sci. 2022, 23, 3780. [Google Scholar] [CrossRef]
  11. Ekins, S.; Puhl, A.C.; Zorn, K.M.; Lane, T.R.; Russo, D.P.; Klein, J.J.; Hickey, A.J.; Clark, A.M. Exploiting machine learning for end-to-end drug discovery and development. Nat. Mater. 2019, 18, 435–441. [Google Scholar] [CrossRef] [PubMed]
  12. Wang, L.; Zhou, Y.; Chen, Q. Ammvf-dti: A novel model predicting drug–target interactions based on attention mechanism and multi-view fusion. Int. J. Mol. Sci. 2023, 24, 14142. [Google Scholar] [CrossRef] [PubMed]
  13. Nguyen, T.; Le, H.; Quinn, T.P.; Nguyen, T.; Le, T.D.; Venkatesh, S. GraphDTA: Predicting drug–target binding affinity with graph neural networks. Bioinformatics 2021, 37, 1140–1147. [Google Scholar] [CrossRef] [PubMed]
  14. Bai, P.; Miljković, F.; John, B.; Lu, H. Interpretable bilinear attention network with domain adaptation improves drug–target prediction. Nat. Mach. Intell. 2023, 5, 126–136. [Google Scholar] [CrossRef]
  15. Zeng, X.; Chen, W.; Lei, B. CAT-DTI: Cross-attention and Transformer network with domain adaptation for drug-target interaction prediction. BMC Bioinform. 2024, 25, 141. [Google Scholar] [CrossRef]
  16. Lee, I.; Keum, J.; Nam, H. DeepConv-DTI: Prediction of drug-target interactions via deep learning with convolution on protein sequences. PLoS Comput. Biol. 2019, 15, e1007129. [Google Scholar] [CrossRef]
  17. Zhu, X.; Liu, J.; Zhang, J.; Yang, Z.; Yang, F.; Zhang, X. FingerDTA: A fingerprint-embedding framework for drug-target binding affinity prediction. Big Data Min. Anal. 2022, 6, 1–10. [Google Scholar] [CrossRef]
  18. Huang, K.; Xiao, C.; Glass, L.M.; Sun, J. MolTrans: Molecular interaction transformer for drug–target interaction prediction. Bioinformatics 2021, 37, 830–836. [Google Scholar] [CrossRef]
  19. Liu, S.; Wang, Y.; Deng, Y.; He, L.; Shao, B.; Yin, J.; Zheng, N.; Liu, T.Y.; Wang, T. Improved drug–target interaction prediction with intermolecular graph transformer. Briefings Bioinform. 2022, 23, bbac162. [Google Scholar] [CrossRef]
  20. Ding, Y.; Tang, J.; Guo, F. Identification of drug-target interactions via multiple information integration. Inf. Sci. 2017, 418, 546–560. [Google Scholar] [CrossRef]
  21. Li, Y.; Huang, Y.A.; You, Z.H.; Li, L.P.; Wang, Z. Drug-target interaction prediction based on drug fingerprint information and protein sequence. Molecules 2019, 24, 2999. [Google Scholar] [CrossRef] [PubMed]
  22. Rogers, D.; Hahn, M. Extended-connectivity fingerprints. J. Chem. Inf. Model. 2010, 50, 742–754. [Google Scholar] [CrossRef] [PubMed]
  23. Yang, J.; Cai, Y.; Zhao, K.; Xie, H.; Chen, X. Concepts and applications of chemical fingerprint for hit and lead screening. Drug Discov. Today 2022, 27, 103356. [Google Scholar] [CrossRef] [PubMed]
  24. Duvenaud, D.K.; Maclaurin, D.; Iparraguirre, J.; Bombarell, R.; Hirzel, T.; Aspuru-Guzik, A.; Adams, R.P. Convolutional networks on graphs for learning molecular fingerprints. Adv. Neural Inf. Process. Syst. 2015, 28, 2224–2232. [Google Scholar]
  25. Cui, F.; Zhang, Z.; Zou, Q. Sequence representation approaches for sequence-based protein prediction tasks that use deep learning. Briefings Funct. Genom. 2021, 20, 61–73. [Google Scholar] [CrossRef]
  26. Wei, W.; Wang, Z.; Mao, X.; Zhou, G.; Zhou, P.; Jiang, S. Position-aware self-attention based neural sequence labeling. Pattern Recognit. 2021, 110, 107636. [Google Scholar] [CrossRef]
  27. Feldmann, C.W.; Sieg, J.; Mathea, M. Analysis of uncertainty of neural fingerprint-based models. Faraday Discuss 2024. [Google Scholar] [CrossRef]
  28. Joshy, A.; Kasyap, G.C.; Reddy, P.D.; Anjusha, I.T.; Nazeer, K.A.A. Drug Target Interaction Prediction using Graph Convo-lution based Neural Fingerprinting. In Proceedings of the 2022 IEEE 19th India Council International Conference (INDICON), Kochi, India, 24–26 November 2022; pp. 1–6. [Google Scholar]
  29. Bian, J.; Lu, H.; Dong, G.; Wang, G. Hierarchical multimodal self-attention-based graph neural network for DTI predic-tion. Briefings Bioinform. 2024, 25, bbae293. [Google Scholar] [CrossRef]
  30. Iyer, R.; Fetterly, G.; Lugade, A.; Thanavala, Y. Sorafenib: A clinical and pharmacologic review. Expert Opin. Pharmacother. 2010, 11, 1943–1955. [Google Scholar] [CrossRef]
  31. Southan, C.; Sharman, J.L.; Benson, H.E.; Faccenda, E.; Pawson, A.J.; Alexander, S.P.; Buneman, O.P.; Davenport, A.P.; McGrath, J.C.; Peters, J.A.; et al. The IUPHAR/BPS Guide to PHARMACOLOGY in 2016: Towards curated quantitative interactions between 1300 protein targets and 6000 ligands. Nucleic Acids Res. 2016, 44, D1054–D1068. [Google Scholar] [CrossRef]
  32. Schöffski, P.; Dumez, H.; Clement, P.; Hoeben, A.; Prenen, H.; Wolter, P.; Joniau, S.; Roskams, T.; Van Poppel, H. Emerging role of tyrosine kinase inhibitors in the treatment of advanced renal cell cancer: A review. Ann. Oncol. 2006, 17, 1185–1196. [Google Scholar] [CrossRef] [PubMed]
  33. Matsui, J.; Yamamoto, Y.; Funahashi, Y.; Tsuruoka, A.; Watanabe, T.; Wakabayashi, T.; Uenaka, T.; Asada, M. E7080, a novel inhibitor that targets multiple kinases, has potent antitumor activities against stem cell factor producing human small cell lung cancer H146, based on angiogenesis inhibition. Int. J. Cancer 2008, 122, 664–671. [Google Scholar] [CrossRef] [PubMed]
  34. Kim, H.; Lim, S.W.; Kim, S.; Kim, J.W.; Chang, Y.H.; Carroll, B.J.; Kim, D.K. Monoamine transporter gene polymorphisms and antidepressant response in Koreans with late-life depression. JAMA 2006, 296, 1609–1618. [Google Scholar] [CrossRef] [PubMed]
  35. Vaishnavi, S.N.; Nemeroff, C.B.; Plott, S.J.; Rao, S.G.; Kranzler, J.; Owens, M.J. Milnacipran: A comparative analysis of human monoamine uptake and transporter binding affinity. Biol. Psychiatry 2004, 55, 320–322. [Google Scholar] [CrossRef]
  36. Chen, X.; Ji, Z.L.; Chen, Y.Z. TTD: Therapeutic target database. Nucleic Acids Res. 2002, 30, 412–415. [Google Scholar] [CrossRef]
  37. Knox, C.; Wilson, M.; Klinger, C.M.; Franklin, M.; Oler, E.; Wilson, A.; Pon, A.; Cox, J.; Chin, N.E.; Strawbridge, S.A.; et al. DrugBank 6.0: The DrugBank Knowledgebase for 2024. Nucleic Acids Res. 2023, 52, D1265–D1275. [Google Scholar] [CrossRef]
  38. Huang, Y.; Dong, D.; Zhang, W.; Wang, R.; Lin, Y.-C.; Zuo, H.; Huang, H.-Y.; Huang, H.-D. DrugRepoBank: A comprehensive database and discovery platform for accelerating drug repositioning. Database 2024, 2024, baae051. [Google Scholar] [CrossRef]
  39. Bank, P.D. Protein data bank. Nat. New Biol. 1971, 233, 10–1038. [Google Scholar]
  40. Vilar, S.; Cozza, G.; Moro, S. Medicinal chemistry and the molecular operating environment (MOE): Application of QSAR and molecular docking to drug discovery. Curr. Top. Med. Chem. 2008, 8, 1555–1572. [Google Scholar] [CrossRef]
  41. Gilson, M.K.; Liu, T.; Baitaluk, M.; Nicola, G.; Hwang, L.; Chong, J. BindingDB in 2015: A public database for medicinal chemistry, computational chemistry and systems pharmacology. Nucleic Acids Res. 2016, 44, D1045–D1053. [Google Scholar] [CrossRef]
  42. Zitnik, M.; Sosic, R.; Leskovec, J. BioSNAP Datasets: Stanford Biomedical Network Dataset Collection. 2018. Available online: http://snap.stanford.edu/biodata (accessed on 20 October 2024).
  43. Wishart, D.S.; Knox, C.; Guo, A.C.; Cheng, D.; Shrivastava, S.; Tzur, D.; Gautam, B.; Hassanali, M. DrugBank: A knowledgebase for drugs, drug actions and drug targets. Nucleic Acids Res. 2008, 36, D901–D906. [Google Scholar] [CrossRef] [PubMed]
  44. Liu, H.; Sun, J.; Guan, J.; Zheng, J.; Zhou, S. Improving compound–protein interaction prediction by building up highly credible negative samples. Bioinformatics 2015, 31, i221–i229. [Google Scholar] [CrossRef] [PubMed]
  45. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  46. Ho, T.K. Random decision forests. In Proceedings of the 3rd International Conference on Document Analysis and Recognition, Montreal, QC, Canada, 14–16 August 1995; Volume 1, pp. 278–282. [Google Scholar]
  47. Chen, L.; Fan, Z.; Chang, J.; Yang, R.; Hou, H.; Guo, H.; Zhang, Y.; Yang, T.; Zhou, C.; Sui, Q.; et al. Sequence-based drug design as a concept in computational drug design. Nat. Commun. 2023, 14, 4217. [Google Scholar] [CrossRef]
  48. Weininger, D. SMILES, a chemical language and information system. 1. Introduction to methodology and encoding rules. J. Chem. Inf. Comput. Sci. 1988, 28, 31–36. [Google Scholar] [CrossRef]
  49. Younesi, A.; Ansari, M.; Fazli, M.; Ejlali, A.; Shafique, M.; Henkel, J. A Comprehensive Survey of Convolutions in Deep Learning: Applications, Challenges, and Future Trends. IEEE Access 2024, 12, 41180–41218. [Google Scholar] [CrossRef]
  50. Pan, X.; Ge, C.; Lu, R.; Song, S.; Chen, G.; Huang, Z.; Huang, G. On the integration of self-attention and convolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 815–825. [Google Scholar]
  51. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  52. Er, M.J.; Zhang, Y.; Wang, N.; Pratama, M. Attention pooling-based convolutional neural network for sentence modelling. Inf. Sci. 2016, 373, 388–403. [Google Scholar] [CrossRef]
  53. Mao, A.; Mohri, M.; Zhong, Y. Cross-entropy loss functions: Theoretical analysis and applications. In Proceedings of the International Conference on Machine Learning, PMLR, Honolulu, HI, USA, 23–29 July 2023; pp. 23803–23828. [Google Scholar]
  54. Demir-Kavuk, O.; Kamada, M.; Akutsu, T.; Knapp, E.W. Prediction using step-wise L1, L2 regularization and feature selection for small data sets with large number of features. BMC Bioinform. 2011, 12, 1–10. [Google Scholar] [CrossRef]
Figure 1. Ablation study on the Human and BioSNAP datasets.
Figure 1. Ablation study on the Human and BioSNAP datasets.
Ijms 25 11818 g001
Figure 2. (a) The 2D visualization result of mycophenolic acids obtained from NFSA-DTI. The orange highlights indicate possible local binding sites, with darker color and larger area indicating greater likelihood. (b,c) The 2D and 3D diagrams of the interaction between mycophenolic acid and inosine monophosphate dehydrogenase from the PDB online database, drawn by the software Molecular Operating Environment (MOE 2019.0102) [40].
Figure 2. (a) The 2D visualization result of mycophenolic acids obtained from NFSA-DTI. The orange highlights indicate possible local binding sites, with darker color and larger area indicating greater likelihood. (b,c) The 2D and 3D diagrams of the interaction between mycophenolic acid and inosine monophosphate dehydrogenase from the PDB online database, drawn by the software Molecular Operating Environment (MOE 2019.0102) [40].
Ijms 25 11818 g002
Figure 3. Learning curves of NFSA-DTI when changing some hyperparameters on the validation set of the BindingDB dataset.
Figure 3. Learning curves of NFSA-DTI when changing some hyperparameters on the validation set of the BindingDB dataset.
Ijms 25 11818 g003
Figure 4. (A) The framework of NFSA-DTI. It includes four stages. (I) The target protein’s amino acid sequence is transformed into a feature matrix, while the drug molecule’s SMILES is processed into graph structure data. (II) The 3-layer CNN processes the two-dimensional feature matrix and obtains the protein representation after the self-attention enhancing unit. Correspondingly, the 3-layer NFGNN processes the graph structure data and obtains the drug representation. (III) Interactions between protein and drug representations are computed via a bilinear attention mechanism, thereby producing a bilinear attention map. (IV) The joint representation obtained after pooling will be input to the fully connected layer for computing the prediction score p. (B) The framework of ESACM. It comprises three 1D convolutional layers, the corresponding 1D batch normalization layers, a self-attention enhancing unit, and a linear layer. The self-attention enhancing unit consists of two stages. (Stage I) Query, key, and value are computed based on the input matrix, in conjunction with their respective weight matrices. (Stage II) The similarity between query and key is evaluated to derive the attention weights. Subsequently, the output is generated through a weighted summation of the values, utilizing the computed attention weights. (C) The flowchart of NFGNN. Firstly, the target node and neighbor nodes surrounding the target node in the molecular graph are integrated and encoded into numerical features. Similar operations are then performed on subsequent nodes to obtain the final neural fingerprint. Subsequently, after the message passing mechanism has completed one iteration of the graph, the neural fingerprint serves as a fixed auxiliary input for updating the graph after each NFlayer.
Figure 4. (A) The framework of NFSA-DTI. It includes four stages. (I) The target protein’s amino acid sequence is transformed into a feature matrix, while the drug molecule’s SMILES is processed into graph structure data. (II) The 3-layer CNN processes the two-dimensional feature matrix and obtains the protein representation after the self-attention enhancing unit. Correspondingly, the 3-layer NFGNN processes the graph structure data and obtains the drug representation. (III) Interactions between protein and drug representations are computed via a bilinear attention mechanism, thereby producing a bilinear attention map. (IV) The joint representation obtained after pooling will be input to the fully connected layer for computing the prediction score p. (B) The framework of ESACM. It comprises three 1D convolutional layers, the corresponding 1D batch normalization layers, a self-attention enhancing unit, and a linear layer. The self-attention enhancing unit consists of two stages. (Stage I) Query, key, and value are computed based on the input matrix, in conjunction with their respective weight matrices. (Stage II) The similarity between query and key is evaluated to derive the attention weights. Subsequently, the output is generated through a weighted summation of the values, utilizing the computed attention weights. (C) The flowchart of NFGNN. Firstly, the target node and neighbor nodes surrounding the target node in the molecular graph are integrated and encoded into numerical features. Similar operations are then performed on subsequent nodes to obtain the final neural fingerprint. Subsequently, after the message passing mechanism has completed one iteration of the graph, the neural fingerprint serves as a fixed auxiliary input for updating the graph after each NFlayer.
Ijms 25 11818 g004
Figure 5. The flowchart of the bilinear attention network. This module consists of two steps. (Step 1) A bilinear interaction matrix is derived through the computation of protein and drug representations. (Step 2) Subsequently, a joint representation is obtained by employing bilinear pooling and attention pooling.
Figure 5. The flowchart of the bilinear attention network. This module consists of two steps. (Step 1) A bilinear interaction matrix is derived through the computation of protein and drug representations. (Step 2) Subsequently, a joint representation is obtained by employing bilinear pooling and attention pooling.
Ijms 25 11818 g005
Table 1. Performance comparison with baseline models on the BindingDB, BioSNAP, and Human dataset (Best).
Table 1. Performance comparison with baseline models on the BindingDB, BioSNAP, and Human dataset (Best).
DatasetsModelAUROCAUPRCAccuracySensitivitySpecificity
BindingDBSVM0.9390.9280.8250.7810.866
RF0.9420.9210.8800.8750.892
DeepConv-DTI0.9450.9250.8820.8730.894
GraphDTA0.9510.9340.8880.8820.897
MolTrans0.9520.9360.8870.8770.902
DrugBAN0.9600.9480.9040.9000.908
CAT-DTI0.9600.9470.8960.8840.913
NFSA-DTI0.9650.9570.9070.9080.906
BioSNAPSVM0.8620.8640.7770.7110.841
RF0.8600.8860.8040.8230.786
DeepConv-DTI0.8860.8900.8050.7600.851
GraphDTA0.8870.8900.8000.7450.854
MolTrans0.8950.8970.8250.8180.831
DrugBAN0.9030.9020.8340.8200.847
CAT-DTI0.9090.9070.8360.8250.847
NFSA-DTI0.9090.9090.8390.8190.858
HumanSVM0.9130.9050.8380.7820.830
RF0.9390.9270.8660.8330.893
DeepConv-DTI0.9780.9820.8780.8300.938
GraphDTA0.9650.9550.9080.9120.904
MolTrans0.9810.9760.9410.9490.939
DrugBAN0.9810.9690.9400.9380.941
CAT-DTI0.9830.9760.9420.9290.957
NFSA-DTI0.9880.9840.9450.9440.955
Table 2. The predicted top 10 drug candidates for Vascular Endothelial Growth Factor Receptor 2 (P35968) in the newly constructed dataset.
Table 2. The predicted top 10 drug candidates for Vascular Endothelial Growth Factor Receptor 2 (P35968) in the newly constructed dataset.
RankDrug NameDrugBank IDEvidence
1SorafenibDB00398Iyer et al. [30]
2RegorafenibDB08896Southan et al. [31]
32-AminobenzimidazoleDB06938Unknown
41-NaphthalenecarboxamideDB07274Unknown
5PonatinibDB08901Unknown
6SunitinibDB01268Schoffski et al. [32]
7Tyrosine Kinase-IN-1DB05014Unknown
8LenvatinibDB09078Matsui et al. [33]
9FostamatinibDB12010Unknown
10RAF265DB05984Southan et al. [31]
Table 3. The predicted top 10 target protein candidates for Nortriptyline (DB00540) in the newly constructed dataset.
Table 3. The predicted top 10 target protein candidates for Nortriptyline (DB00540) in the newly constructed dataset.
RankProtein NameUniprot IDEvidence
1Sodium-dependent noradrenaline transporterP23975Kim et al. [34]
2Alpha-2A adrenergic receptorP08913Unknown
35-hydroxytryptamine receptor 2AP28223Southan et al. [31]
45-hydroxytryptamine receptor 1AP08908Southan et al. [31]
55-hydroxytryptamine receptor 1CP08909Southan et al. [31]
65-hydroxytryptamine receptor 2CP28335Southan et al. [31]
7Alpha-1B adrenergic receptorP35368Unknown
8Sodium-dependent serotonin transporterP31645Vaishnavi et al. [35]
9Beta-1 adrenergic receptorP08588Unknown
10Muscarinic acetylcholine receptor M1P11229Unknown
Table 4. Experimental datasets statistics.
Table 4. Experimental datasets statistics.
DatasetsDrugsProteinsInteractions*P2N
BindingDB14,643262349,1990.725
BioSNAP4510218127,4641.014
Human2726200167281
Table 5. Hyperparameters configuration.
Table 5. Hyperparameters configuration.
ModuleHyperparametersValue
ESACMInitial amino acid embedding dimension128
Kernel size[3, 6, 9]
Number of filters[128, 128, 128]
Heads of self-attention2
NFGNNInitial atom embedding dimension128
Hidden node dimensions[128, 128, 128]
Bilinear attention networkHeads of bilinear attention2
Bilinear embedding dimension768
Attention pooling window size3
Attention pooling stride3
Fully connected decoderNumber of hidden neurons512
OptimizerLearning rate5 × 10−5
Epoch100
Mini-batchBatch size64
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, F.; Xu, H.; Cui, P.; Li, S.; Wang, H.; Wu, Z. NFSA-DTI: A Novel Drug–Target Interaction Prediction Model Using Neural Fingerprint and Self-Attention Mechanism. Int. J. Mol. Sci. 2024, 25, 11818. https://doi.org/10.3390/ijms252111818

AMA Style

Liu F, Xu H, Cui P, Li S, Wang H, Wu Z. NFSA-DTI: A Novel Drug–Target Interaction Prediction Model Using Neural Fingerprint and Self-Attention Mechanism. International Journal of Molecular Sciences. 2024; 25(21):11818. https://doi.org/10.3390/ijms252111818

Chicago/Turabian Style

Liu, Feiyang, Huang Xu, Peng Cui, Shuo Li, Hongbo Wang, and Ziye Wu. 2024. "NFSA-DTI: A Novel Drug–Target Interaction Prediction Model Using Neural Fingerprint and Self-Attention Mechanism" International Journal of Molecular Sciences 25, no. 21: 11818. https://doi.org/10.3390/ijms252111818

APA Style

Liu, F., Xu, H., Cui, P., Li, S., Wang, H., & Wu, Z. (2024). NFSA-DTI: A Novel Drug–Target Interaction Prediction Model Using Neural Fingerprint and Self-Attention Mechanism. International Journal of Molecular Sciences, 25(21), 11818. https://doi.org/10.3390/ijms252111818

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop