Next Article in Journal
Long-Term In Vitro Assessment of Biodegradable Radiopaque Composites for Fiducial Marker Fabrication
Next Article in Special Issue
Ligand-Based Drug Design of Novel Antimicrobials against Staphylococcus aureus by Targeting Bacterial Transcription
Previous Article in Journal
Chemical Element Profiling in the Sera and Brain of Bipolar Disorders Patients and Healthy Controls
Previous Article in Special Issue
In Vitro and In Silico Analysis of New n-Butyl and Isobutyl Quinoxaline-7-carboxylate 1,4-di-N-oxide Derivatives against Trypanosoma cruzi as Trypanothione Reductase Inhibitors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Docking and Machine Learning Approach to Enhance the Performance of Virtual Screening Carried out on Protein–Protein Interfaces

NeuroDiderot Department, Université de Paris, Inserm UMR 1141, Robert-Debré Hospital, 75019 Paris, France
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2022, 23(22), 14364; https://doi.org/10.3390/ijms232214364
Submission received: 28 September 2022 / Revised: 11 November 2022 / Accepted: 16 November 2022 / Published: 18 November 2022
(This article belongs to the Special Issue New Avenues in Molecular Docking for Drug Design 2022)

Abstract

:
The modulation of protein–protein interactions (PPIs) by small chemical compounds is challenging. PPIs play a critical role in most cellular processes and are involved in numerous disease pathways. As such, novel strategies that assist the design of PPI inhibitors are of major importance. We previously reported that the knowledge-based DLIGAND2 scoring tool was the best-rescoring function for improving receptor-based virtual screening (VS) performed with the Surflex docking engine applied to several PPI targets with experimentally known active and inactive compounds. Here, we extend our investigation by assessing the vs. potential of other types of scoring functions with an emphasis on docking-pose derived solvent accessible surface area (SASA) descriptors, with or without the use of machine learning (ML) classifiers. First, we explored rescoring strategies of Surflex-generated docking poses with five GOLD scoring functions (GoldScore, ChemScore, ASP, ChemPLP, ChemScore with Receptor Depth Scaling) and with consensus scoring. The top-ranked poses were post-processed to derive a set of protein and ligand SASA descriptors in the bound and unbound states, which were combined to derive descriptors of the docked protein-ligand complexes. Further, eight ML models (tree, bagged forest, random forest, Bayesian, support vector machine, logistic regression, neural network, and neural network with bagging) were trained using the derivatized SASA descriptors and validated on test sets. The results show that many SASA descriptors are better than Surflex and GOLD scoring functions in terms of overall performance and early recovery success on the used dataset. The ML models were superior to all scoring functions and rescoring approaches for most targets yielding up to a seven-fold increase in enrichment factors at 1% of the screened collections. In particular, the neural networks and random forest-based ML emerged as the best techniques for this PPI dataset, making them robust and attractive vs. tools for hit-finding efforts. The presented results suggest that exploring further docking-pose derived SASA descriptors could be valuable for structure-based virtual screening projects, and in the present case, to assist the rational design of small-molecule PPI inhibitors.

1. Introduction

Virtual screening (VS) is one popular set of techniques to extract a list of potentially bioactive molecules from input chemical libraries [1,2,3,4,5,6,7]. Structure-based methods are of high interest to investigate novel mechanisms, such as protein–protein interactions, even more so in the era of AI-powered protein structures predictions [8,9,10,11] and major advances in the field of structural biology, such as Cryo-EM [12,13,14]. Structure-based virtual screening (SBVS) has been shown efficient in suggesting hit compounds on numerous targets, but the approach is not without pitfalls [15,16,17,18,19]. An important problem in virtual screening is that the methods are usually target-dependent, and as such, one method cannot fit all [15,17,18,19,20,21]. In SBVS studies, the compounds are first positioned in a binding pocket and then scored. Several docking programs have been shown proficient in reasonably sampling the binding site space and in generating reliable poses [20], but a key problem is scoring, resulting in a high number of inactive compounds ranked high in the scoring list or of active molecules lost. There are many sources of errors in scoring [15,16,17,18,19], including the difficulties of implementing complex energy terms in fast scoring functions and the obviously related challenges of taking into consideration water molecules and protein flexibility, among others. Several strategies have been proposed to try to improve the process, such as the use of more sophisticated CPU/GPU intensive energy computations, but the advantages are not obvious and are also target-dependent [15]. Molecular interaction fingerprints, machine learning (ML) scoring functions, target-specific scoring functions, and consensus scoring represent other approaches that have been shown valuable in some circumstances [22,23,24,25,26,27,28,29,30,31,32,33,34].
Protein–protein interactions (PPIs) regulate diverse types of biological activities both in healthy and disease states [35]. It is estimated that the human interactome comprises about 130,000–650,000 types of PPI [36,37]. The crucial importance of PPIs makes them a rich source of putative targets for the development of a new generation of therapeutics or for the design of chemical probes that could be used to validate further the importance of some specific PPIs. However, the nature of the PPI interfaces is in general significantly different from that of classical drug targets, which have well-defined pockets allowing the tight binding of small molecules [38]. The PPI binding interface is usually large (~1500–3000 Å2) and flat (often exposed to solvent) [39], posing substantial problems in ligand discovery using structure-based approaches. Our last study revealed DLIGAND2 as a best-rescoring function that significantly improved the vs. performance of Surflex docking for various PPI targets. DLIGAND2 is a knowledge-based scoring function that predicts protein–ligand binding affinity based on a distance-scaled, finite, ideal-gas reference (DFIRE) state [40]. Here, we extend our previous research by assessing the vs. potential of rescoring using GOLD scoring functions (GoldScore, ChemScore, Astex Statistical Potential (ASP), ChemPLP, ChemScore with Receptor Depth Scaling (RDS)). In addition, the poses were also scored using a consensus method involving all scoring functions. The best-predicted poses sorted on the Surflex docking score were further post-processed to derive 3D SASA descriptors of both the targets and the ligands in their bound and unbound states. The protein and ligand SASA descriptors were combined to derive a set of protein–ligand SASA descriptors. These descriptors were independently assessed to measure any possible impact on vs. metrics. Finally, these descriptors were employed for building eight diverse ML models: tree, bagged forest, random forest, Bayesian, support vector machine (SVM), logistic regression, neural network, and neural network with bagging. Several studies have been performed by exploiting chemical fingerprints to build ML-based quantitative structure–activity relationship (QSAR) models to separate true inhibitors of PPIs from inactive molecules [41,42,43,44]. These studies essentially aimed at the investigation of the physicochemical property ranges of compounds acting as inhibitors of PPIs. The identified properties were then used to build statistical models enabling the generation of compound collections likely enriched in inhibitors of PPIs [45,46,47]. However, the potential and effectiveness of descriptors calculated from the experimental 3D structure of a protein–ligand complex or of docking pose-generated 3D descriptors have not been fully investigated so far either for rescoring and ranking PPI compounds or for building classification models. The primary assumption in using these topological 3D descriptors to rank the poses is that the docking programs are capable of correctly sampling the binding site pockets and generating reliable poses, but they are not always accurate at scoring them for a wide range of protein targets. Here, we are interested in the evaluation of several post-processing strategies. First, we investigated the impact of using different types of scoring functions and a consensus scoring method on several PPI targets. Then, we investigated the potential benefit of using SASA data to rescore the molecules. In this case, we used the best docking poses generated by a fast method available in Surflex (assumed to be reasonable) and grades them using the derivatized SASA score. Then, we explored further the SASA descriptors, through various ML approaches, with the aims of generating some predictive classification models and identifying key SASA descriptors capable of discriminating between true active and true inactive ligands. The vs. performance on different targets was assessed by computing receiver operating characteristics (ROC) graphs and calculating the area under the curve (AUC) and the early recognition metrics (BEDROC, enrichment factors at 1% and 5%) to determine the ability of different rescoring methods to rank the active compounds in the top positions. The performance of ML models was estimated by calculating AUC, sensitivity, specificity, precision, and concordance. The results indicate that among the different GOLD scoring functions, GoldScore and ASP were found valuable for several targets while many SASA descriptors performed equivalently to or better than the default Surflex and GOLD scoring functions for most targets. Among the ML models, neural network and random forest classification models yielded superior vs. results for most targets. The present study provides valuable insights regarding rescoring strategies, and ML methods on PPIs. The presented protocols could assist the identification of novel drug-like orthosteric PPI inhibitors. This can be of significant interest with applications on novel targets acting in different diseases or to replace some monoclonal antibodies or peptides inhibiting already known targets.

2. Results

The workflow describing the hybrid computational approach developed in this study by employing docking-based VS, rescoring, and ML is shown in Figure S1. The ten PPI targets used here include the Bromodomain Adjacent to Zinc Finger Domain 2B (BAZ2B), Apoptosis regulator Bcl-2; Apoptosis regulator Bcl-xL, BRD4 bromodomain 1 BRD4-1, CREB-binding protein (CREBBP), HIV Integrase (HIV IN), Inhibitor of apoptosis protein 3 (XIAP), Induced myeloid leukemia cell differentiation protein Mcl-1, and E3 ubiquitin-protein ligase Mdm2, and Menin (Table 1, Figure S2). The Ephrin type-A receptor 4 (EphA4) was excluded from this work as there were not sufficient compounds to build ML models. The vs. performance of rescoring and ML models was assessed on two protein structures per target, which were selected based on the diversity of the binding sites (Figure S2). The target compounds were collected from the ChEMBL [48] and PubChem [49] databases. The chemical space of the PPI datasets was analyzed by performing Principal Component Analysis (PCA) using simple physicochemical properties of the compounds (Figure S3). For more details on the methodology used for selecting the protein structures, curation of target datasets, and PCA analysis, see ref. [50].

2.1. Virtual Screening Performance of GOLD Scoring Functions

We observed in our previous study that the fast “pscreen” Surflex docking approach, on PPI targets, displayed variable screening performances with excellent AUC values (>0.8) for 3 targets (Bcl-2, Bcl-xl, XIAP); moderate AUCs (0.6–0.7) for four targets (BRD4-1, CREBBP, HIV IN, Mdm2); and poor AUC values (0.4–0.6) for four targets (BAZ2B, Mcl-1, Menin and EphA4) [50]. We now wanted to investigate the performance of other types of re-scoring functions on this dataset. We decided to use the GOLD package for rescoring purposes as it offers different types of scoring functions. For example, the GoldScore function combines force-field terms with empirical terms to account for some of the deficiencies in pure force-field-based scoring functions [51]. ASP is a knowledge-based scoring function that uses information about the frequency of interaction between ligand and protein atoms. It is gathered by analyzing existing ligand-protein structures in the PDB and this information is then used to generate the statistical potentials. ChemPLP and ChemScore are pure empirical scoring functions that estimate the binding affinity of a complex based on a set of weighted energy terms whose coefficients are determined by fitting the binding affinity data of a training set of protein–ligand complexes with known 3D structures, using least squares fitting (see Methods for more detail on GOLD scoring functions). Table 2 shows the comparison of vs. performance between Surflex and different GOLD scoring functions (GoldScore, ChemScore, ASP, ChemPLP, ChemScore RDS) in terms of AUC and BEDROC for the different PPI targets. The AUC calculations were performed after sorting the global pose space and retaining the best pose for each compound after rescoring with GOLD scoring functions and via a consensus approach. Table S1 shows the corresponding early enrichment values for the targets at 1% and 5% levels for the best-predicted poses obtained from rescoring the global pose space. The rescoring of the different data sets revealed that among all the GOLD scoring functions, ASP and GoldScore displayed the best results by yielding higher AUC values for six targets (with a total of 11 protein structures) and seven targets (10 protein structures), respectively, as compared to Surflex. Importantly, ASP generated significantly improved and highest AUC values for four protein structures (Bcl-2: 2O21; BRD4-1: 5D3L, 5KU3; CREBBP: 5EIC), while GoldScore produced the best AUC values for three structures (Mcl-1: 5MES, Mdm2: 4ODF, 4ZFI). ChemPLP and consensus were the next good performing scoring functions generating higher AUC values than Surflex for eight and nine protein structures respectively. The consensus approach generated an AUC of 0.66 in 5FC4 (Mcl-1), which was the highest among all scoring functions. The BEDROC calculations showed that ChemPLP was the best scoring function producing higher values for 12 target protein structures, followed by GoldScore and consensus scoring for 11 structures as compared to Surflex. ASP and ChemScore generated higher BEDROC values for nine protein structures. GoldScore produced the BEDROC values with a difference greater than 0.05 (ΔBEDROC) for two structures (Mcl-1: 5FC4, 5MES) as compared to Surflex and with a ΔBEDROC of >0.1 for three structures (BRD4-1: 5D3L, 5KU3, Mdm2: 4ZFI), indicating good early recovery of the actives. Similarly, ChemPLP produced ΔBEDROC > 0.05 for two target structures (Bcl-2: 2O21, Mcl-1: 5FC4) and ΔBEDROC > 0.1 for three structures (Bcl-xL: 3WIZ, BRD4-1). Consensus scoring produced a ΔBEDROC of >0.05 for three structures (Bcl-xL: 3WIZ, Mcl-1: 5FC4, CREBBP: 5EIC) and ΔBEDROC > 0.1 for two structures (BRD4-1). Overall, all rescoring functions produced a ΔBEDROC > 0.1 in BRD4-1 indicating a significantly improved performance compared to Surflex on this PPI dataset. In the case of 5KU3, the rescoring functions produced more than a two-fold improvement in the BEDROC score.
Consistent with the preceding observations, ChemPLP and consensus scoring produced high enrichment values in six targets with a total of 12 and 10 protein structures at a 1% level, respectively, as compared to Surflex. Whereas, GoldScore and ASP produced a high EF1% for 6 targets (nine protein structures) and five targets (nine structures), respectively. The results at the 5% level were much better with GoldScore, ChemPLP, and consensus scoring. They all produced EF values better than Surflex in seven targets (ten protein structures for GoldScore, twelve for ChemPLP, and eleven for consensus). Whereas ASP and ChemScore generated high enrichment values for four targets (eight protein structures) and five targets (eight structures), respectively. There were some target proteins for which the enrichment factors were more than three times higher as compared with those obtained with Surflex. For example, Menin (6B41) has a ≥3-fold increase in its EF1 % when rescored with GoldScore, ASP, and ChemPLP. In Mcl-1 (5MES), EF 1% is more than two times using GoldScore. Similarly, in the case of 5KU3 (BRD4-1), GoldScore, ASP, and consensus scoring produced EF 5% > 3-fold, while in the structure 5D3L, EF1% was >2-fold using ChemScore RDS. Overall, it can be summarized that rescoring the Surflex-generated poses with GOLD scoring functions can be advantageous for some targets. In particular, GoldScore and ASP generated highly improved performances in targets, such as Mcl-1, Mdm2, BRD4-1, and Bcl-2, while for other targets, the AUCs were comparable to or slightly better than Surflex. Nevertheless, like Surflex, these scoring functions tend to also be target-dependent.

2.2. Post-Docking Derivatization of SASA Descriptors

We showed previously that the relative SASA (rSASA) (see Section 4.2) of the ligand could serve as a robust descriptor for identifying true PPI ligands, which are largely solvent-exposed at the binding interface. The rSASA showed good vs. performance when used independently or in combination with the Surflex scoring function [50]. Encouraged by these results, we set out to explore the capability of several diverse protein, ligand, and protein–ligand SASA descriptors in modulating vs. performance. These descriptors were calculated by measuring the SASA of the receptor–ligand complex (the “bound” surface area), the receptor on its own, and the ligand on its own (the “unbound” surface areas”), and then calculating the difference in surface area between the bound complex and the unbound receptor and ligand (the “delta” surface area). No changes are made to the structures, so relaxation on binding is not included in the calculations. The SASA of the ligand and receptor before and after the receptor–ligand binding can be broken down into different categories based on either pharmacophore types, residue types, or structural features. Following the pharmacophoric approach, the SASA in the specified state of all receptor or ligand atoms within the system is divided individually into six basic types: Hydrophobic (H), Aromatic (Ar), donor (D), Acceptor (Ac), Positive (P), Negative (N). Those atoms not matching any of the types were classified as the type “other” (O). By using this approach, we extracted the seven SASA descriptor subtypes for the ligand and the receptor in the bound and unbound state. We next calculated the delta and relative SASA for each SASA category for the receptor and the ligand. In the end, four global ligand and receptor SASA descriptors were derived: total (bound and unbound), total delta, and relative total (see Methods). In this way, we derived
32 receptor SASA descriptors: total unbound (TRU), total bound (TRB), total delta (ΔTR), relative total (rel. TR), hydrophobic unbound (HRU), hydrophobic bound (HRB), hydrophobic delta (ΔHR), relative hydrophobic (rel. HR), aromatic unbound (ArRU), aromatic bound (ArRB), aromatic delta (ΔArR), relative aromatic (rel. ArR), donor unbound (DRU), donor bound (DRB), donor delta (ΔDR), relative donor (rel. DR), acceptor unbound (AcRU), acceptor bound (AcRB), acceptor delta (ΔAcR), relative acceptor (rel. AcR), positive unbound (PRU), positive bound (PRB), positive delta (ΔPR), relative positive (rel. PR), negative unbound (NRU), negative bound (NRB), negative delta (ΔNR), relative negative (rel. NR), other unbound (ORU), other bound (ORB), other delta (ΔOR), relative other (rel. OR);
32 ligand SASA descriptors: total unbound (TLU), total bound (TLB), total delta (ΔTL), relative total (rel. TL), hydrophobic unbound (HLU), hydrophobic bound (HLB), hydrophobic delta (ΔHL), relative hydrophobic (rel. HL), aromatic unbound (ArLU), aromatic bound (ArLB), aromatic delta (ΔArL), relative aromatic (rel. ArL), donor unbound (DLU), donor bound (DLB), donor delta (ΔDL), relative donor (rel. DL), acceptor unbound (AcLU), acceptor bound (AcLB), acceptor delta (ΔAcL), relative acceptor (rel. AcL), positive unbound (PLU), positive bound (PLB), positive delta (ΔPL), relative positive (rel. PL), negative unbound (NLU), negative bound (NLB), negative delta (ΔNL), relative negative (rel. NL), other unbound (OLU), other bound (OLB), other delta (ΔOL), relative other (rel. OL).
With the protein and ligand SASA descriptors in hand, we derived 16 protein–ligand SASA descriptors for each SASA category: total delta (ΔTRL), relative total (rel. TRL), hydrophobic delta (ΔHRL), relative hydrophobic (rel. HRL), aromatic delta (ΔArRL), relative aromatic (rel. ArRL), donor delta (ΔDRL), relative donor (rel. DRL), acceptor delta (ΔAcRL), relative acceptor (rel. AcRL), positive delta (ΔPRL), positive (rel. PRL), negative delta (ΔNRL), relative negative (rel. NRL), other delta (ΔORL), relative other (rel. ORL) (see methods).

2.3. Virtual Screening Performance of SASA Descriptors

The AUC values calculated for the PPI datasets screened using SASA descriptors are shown as a heat map in Figure 1 and Table S2 in the supporting information. The combined analysis of AUCs for the different descriptors is shown in Figure S4. The percent change in the AUC values, BEDROC scores, and database enrichments EF1%/EF5% are shown in Tables S3–S6, respectively. The results indicate that many post-docking calculated SASA descriptors of proteins, ligands, and protein–ligand complexes generated AUCs that were comparable to or better than the default Surflex scores in many targets. For instance, among the receptor SASA descriptors, TRU, ΔTR, ΔHRU, ΔHR, and ΔArR were the best descriptors that achieved excellent performance (AUC > 0.8) in seven, nine, six, eight, and five protein structures, respectively, and reasonably good performance (AUC: 0.7–0.8) in another three, one, four, two and three target structures, respectively. Bcl-2, Bcl-xL, BRD4-1, XIAP, and Mdm2 were the common test systems in which TRU, ΔTR, ΔHRU, and ΔHR descriptors showed good performance. Whereas, ΔArR showed excellent performance in the aforementioned targets except for XIAP. When compared to the performance of Surflex, TRU, ΔTR, HRU, ΔHR, ΔArR, DRU, ΔDR, AcRU, and ΔAcR generated improved AUC values in 12, 13, 14, 15, 11, 10, 10, 10, and nine protein structures respectively (Table S3). Among the ligand SASA descriptors, TLU, TLB, ΔTL, OLU, OLB, and ΔOL achieved excellent performance (AUC > 0.8) in 8, 6, 8, 6, 6, and 6 protein structures, respectively. Bcl-2, Bcl-xL, BRD4-1, XIAP, and Mdm2 were the common targets in which these ligand descriptors performed very well. However, the performance of OLU, OLB, and ΔOL was moderate in the targets BRD4-1 and XIAP. When considering overall performance, TLU, TLB, ΔTL, HLU, OLU, OLB, and ΔOL were the best descriptors that generated higher AUC values as compared to Surflex in 15, 15, 16, 11, 11, 12, and nine protein target structures, respectively (Table S3). Among the receptor–-ligand SASA descriptors, ΔTRL, ΔHRL, ΔDRL, and ΔORL produced excellent results (AUC > 0.8) in eight, five, six, and six protein structures, respectively. Whereas ΔTRL and ΔHRL showed good performance in another target structure. Bcl-2, Bcl-xL, and XIAP were again the common targets in which ΔTRL, ΔDRL, and ΔORL showed improved performance. ΔTRL also showed high performance in BRD4-1, while ΔHRL showed excellent results in Bcl-2, Bcl-xL, and BRD4-1. Based on the results obtained against all targets, it can be inferred that ΔTRL, ΔHRL, ΔARL, ΔDRL, ΔNRL, and ΔORL were the best descriptors generating higher AUC values as compared to Surflex in 16, 12, 10, 10, nine, and nine protein structures respectively. Based on the average AUCs over the entire dataset, the following trend was observed: TLU > ΔTR > ΔTRL > ΔHR > ΔTL > TRU > TLB > HRU > OLU > ΔHRL > OLB > ΔArR > Surflex > others (Figure S4). Besides the good vs. performance, as evident by the enhanced AUC values (Figure 1 and Table S2), the SASA descriptors-based scoring metrics also exhibited high early enrichments and BEDROC scores in comparison to Surflex. The SASA descriptors TRU, ΔTR, HRU, ΔHR, TLU, TLB, ΔTL, OLU, OLB, ΔTRL, ΔHRL, and ΔArRL generated higher EF 1% values in eight targets (with a total of 12 protein structures), eight targets (11 structures), six (nine structures), eight targets (11 structures), nine targets (14 structures), eight targets (11 structures), eight targets (12 structures), six targets (nine structures), seven targets (10 structures), eight targets (12 structures), seven targets (10 structures), and five targets (nine structures), respectively. There were some target proteins for which the EF1% is 2–6 times higher than those obtained with Surflex. For example, Mcl-1 (5MES) has a ~4-fold increase in its EF1% for TLB, ΔTRL; and >4-fold and >5-fold when scored using TR, and TLU, respectively. Similarly, Menin (6B41) has a >4-fold increase in its EF1% for TR and DR. Whereas, TRU, HR, AcRU, AcR, NR, TLU, TL, OLU, ΔOL, and ΔORL produced increased enrichment (between 3–4%) for the same protein. In the case of BRD4-1 (5KU3), rel. HL showed a ~3-fold increase in EF1% value. TR and ΔTRL produced ~2 times higher EF1% value for Bcl-xL (3INQ). When considering EF5%, TRU showed higher enrichment values in six targets (with a total of 11 protein structures); ΔTR in eight targets (14 protein structures); HRU in six targets (11 structures); ΔHR in eight targets (14 structures); ΔArR in six targets (10 structures); TLU in eight targets (13 structures); TLB in eight targets (12 structures); TL in seven targets (11 structures); OLB in six targets (11 structures); TRL in eight targets (13 structures); and ΔHRL in eight targets (12 structures). There were some target proteins for which the EF5% is up to four times higher than those obtained from Surflex docking. For example, rel. HL showed a four-fold increase of EF5% in BRD4-1 (5KU3). In the case of Bcl-xL (3WIZ), TLU generated EF5% ~2-fold higher as compared to Surflex. TLU and TLB generated more than two-fold higher EF5% in 5MES (Mcl-1). Whereas ΔTR, ΔTR, and ΔTRL produced EF5% > two-times in Menin (6B41). Collectively, among all descriptors, TRU, ΔTR, ΔHR, TLU, TLB, ΔTL, OLU, and ΔTRL were the major SASA scoring metrics that produced higher average EF1% and EF5% values as compared to Surflex (Figure S5). The vs. performance in terms of BEDROC score (α = 20) for the SASA descriptors showed a similar trend as that observed for the early enrichment factors (Figure S6). There were some target proteins for which the BEDROC scores were 2–3 times higher than those obtained from the default Surflex-dock scores. For instance, HLB and rel. HL generated BEDROC score ~3 times higher for BRD4-1 (5KU3) as compared to Surflex. TR, TLU, TLB, and TRL increased the BEDROC value by ~2-fold in Mcl-1 (5MES). While TLU produced ~2 times higher BEDROC score in the case of 5FC4 (Mcl-1). Thus, it can be deduced that the performance of some topological descriptors is superior to Surflex as well as GOLD rescoring functions. This suggests that these 3D descriptors could be used as scoring metrics for screening compound collections after performing docking and for instance a first rescoring post-processing step.

2.4. Building and Validating Machine Learning Models Using SASA Descriptors

To investigate whether it is possible to discriminate the two classes of PPI inhibitors (active: 1 and inactive: 0) using the ML approaches, we built several classification ML models relying on the derivatized SASA descriptors from the docking poses for each PPI target and the corresponding protein structures. The dataset (poses) of each target was split into a training set (70%) and a test set (30%) in a stratified fashion to have the same proportion of class labels in the training and test subsets as the input data set (Table 1). We selected eight binary classifiers, tree, bagged forest, random forest, Bayesian, SVM, logistic regression, neural net, and neural net (with bagging) implemented in BIOVIA pipeline pilot v18.1 software. The models were constructed with default hyperparameters. The model performance was assessed by computing ROC AUC, sensitivity, specificity, precision, and concordance (accuracy) for the training and test sets. Table 3 shows the 10-fold cross-validation ROC AUC values attained for ten PPI targets using eight ML methods. Other model statistics obtained from the cross-validation are provided in Table S7. Based on the results, the neural network (bagging) outperformed all ML techniques by producing the highest cross-validated AUC values in nine out of 10 targets (with a total of 17 protein structures). SVM produced the best AUC values in both structures of BAZ2B and in one structure of CREBBP (5MMG). Both neural network (bagging) and SVM were equally effective in the case of 3WIZ (Bcl-xL) and 5EIC (CREBBP). Similarly, random forest and neural network (bagging) showed the best performance in the case of Bcl-2 (2O21). Taken together, neural net (bagging), random forest, and SVM emerged as the best classifiers by producing AUC > 0.8 in 18 structures; Bayesian and neural network in 14 structures; and tree, bagged forest, logistic regression in 12 structures. Table 4 shows the AUC values obtained by the different models on the test sets. Whereas other statistics obtained on test sets are provided in Tables S8 and S9. Neural network with bagging produced the best AUC values in eight out of 10 targets with a total of 14 structures. The random forest algorithm produced the best AUC values in two targets (Bcl-2 and Menin; four structures). Tree with or without bagging generated the highest AUC value in BRD4-1 (5D3L). The neural network without bagging showed the highest AUC values in Bcl-xL (3INQ) and XIAP (5C3H). With respect to overall performance, neural networks with bagging achieved excellent performance (AUC > 0.8) in 14 structures; random forest in 13 structures; trees with bagging, SVM, and logistic regression, and neural network without bagging in 12 structures each, tree without bagging in 11 structures; and Bayesian in 10 structures. Considering the performances of ML classifiers on the test sets, the different algorithms can be graded in the following order: neural network (bagging) > random forest > bagged forest ~ SVM ~ logistic regression ~ neural network > tree > Bayesian. A similar trend can be deduced when the sensitivity, specificity, precision, and concordance data are compared across the different ML models. For instance, the neural net (bagging) model produced sensitivity > 80% in six targets (with a total of 12 protein structures). SVM, Bayesian, random forest, and neural network produced sensitivity > 0.8 in six targets (11 structures), five targets (10 structures), five targets (nine protein structures), and five targets (nine structures), respectively (Table S8). However, most of the ML models failed in terms of sensitivity in the targets BAZ2B and Menin.

2.5. Enrichment Factors Estimation of Machine Learning Models

All of the ML models were used to calculate the enrichment factors (EF1% and EF5%) for the test sets (Table S10). Consistent with the highly improved AUC values, the ML models also showed high early enrichment values for the test sets in several target proteins. In the case of BAZ2B, the best enrichment in the hit rate for the top-scoring 1% of compounds was 2 for the protein 4XUA produced by neural net (bagging), and for 5E73 SVM produced EF1% of 2.2. For Bcl-2 (2O21), random forest, SVM, and neural networks were equally effective by producing the best EF1% value of 28.72. In the case of 4LVT, these models, including Bayesian, produced the same EF1% value of 28.72. For Bcl-xL (3INQ), neural network (bagging) produced the best EF1% of 52. While in the case of 3WIZ, neural networks and random forest produced the same EF1% value. For BRD4-1 (5D3L), all ML models, except Bayesian, produced the EF1% of 2.162. In comparison, all models produced the same result in the case of 5KU3. For CREBBP, the neural network (bagging) exhibited the best EF1% value of 12.38 and 9.03 in 5EIC and 5MMG, respectively. For HIV IN, all methods exhibited the EF1% value of 2.4 in 4CFD. While in 4CHO, bagged forest, random forest, and neural networks showed the same EF1% value. For XIAP, random forest and neural network (bagging) produced the best EF1% value of 43.934 in 1TFT and 5C3H. For Mcl-1, SVM and neural network (bagging) displayed the best enrichment of 23.805 and 23.57 in 5FC4 and 5MES, respectively. For Mdm2, all models, except Bayesian and logistic regression, exhibited the EF1% value of 2.9. In the case of 4ZFI, random forest, SVM, logistic regression, and neural network (without bagging) produced the same EF1%. For Menin, SVM produced the best EF1% value of 6.68 in both 5DB2 and 6B41. Overall, the best performance was exhibited by the neural network by yielding the best EF1% value in nine targets with a total of 14 structures and an average EF1% of 17.23. The second-best ML method was SVM which produced the best EF1% value in seven targets with a total of 11 structures and an average EF1% of 17.01. Random forest stood in third place by producing the best EF1% value in seven targets with a total of 10 structures and an average EF1% of 16.8. With respect to the enrichment factor at 5%, the neural network (bagging) outperformed other techniques by producing the best EF5% value in all targets with a total of 14 structures and an average EF5% of 8.5. Random forest was the second-best method that generated the high EF5% value in six targets with a total of seven structures and an average EF5% value of 8.3. SVM produced a high EF5% value in four targets (four structures) and an average EF5% of 8.134. Our analysis of additional test statistics (Cohen’s Kappa κ, Matthew’s correlation coefficient MCC, Youden’s Index J, and F1 score) is in agreement with the conclusions drawn here. For example, the Neural network with bagging performed better than other methods based on the aforementioned parameters in most of the targets. Whereas, Bayesian and SVM produced some improvement in the targets BAZZ2B and Menin, in which other methods failed (Figure S7). We also compared the early enrichments of ML models to Surflex and the best-performing receptor, ligand, and receptor–ligand SASA descriptors (Figure S8). The overall mean of EF1% and EF5% for the ML models were much higher than the Surflex and SASA descriptors. For instance, neural network (bagging) models produced EF1% ~7 times better as compared to Surflex in targets Mcl-1 (5FC4) and CERRBP (5EIC). Whereas the early recovery values obtained using SASA descriptors were better than Surflex. These results indicate that the docking scoring function may not be efficient when used solely for performing VS. Rather, the use of docking pose-derived descriptors in combination with ML techniques could be valuable in identifying active PPI ligands. We also analyzed the predictor fields to understand which features matter most and which are of least importance based on the relative importance of each predictor in estimating the model. Figure S9 shows the distribution of descriptor importance values obtained from the 20 neural net (bagging) models trained on different data sets. ΔHR, ΔTL, ΔTR, TRU, ΔTRL, HRU, TLU, TLB, TRB, and ΔHRL descriptors were predicted to be of high importance based on the Neural net models. All these descriptors also showed good vs. performance when assessed independently.

3. Discussion

With the discovery of novel therapeutically important PPIs, there is an urgent need to develop drug candidates modulating these interactions to supplement the currently limited clinical pipeline. Alternatively, small molecules that perturb specifically some PPIs could be valuable to understand the importance of a protein complex in a disease state and thus help de-risk target selection. However, the rational design of selective and potent small molecule inhibitors of PPIs is in general very challenging compared to traditional targets, such as kinases, proteases, or G protein-coupled receptors [52,53,54]. The present study is built on our previous fast docking and rescoring-based vs. investigation performed on inhibitors of PPIs and aims at gaining insights about the pertinence of using hybrid methods that combine docking pose-derived SASA descriptors with scoring functions and/or with ML techniques (i.e., when enough data are available for the target of interest) so as to facilitate the discovery of novel PPI inhibitors. This work is also encouraged by the research performed by Núñez et al. in which they applied a similar strategy to traditional drug targets: adenosine deaminase, and estrogen receptor alpha [55]. In that study, after docking with GOLD and Glide, they post-processed the poses to derive a protein–ligand interaction fingerprint (PLIF) metric. Next, the SASA descriptors were computed for each ligand and its respective protein in their bound/unbound states. Subsequently, a Bayesian model was learned with SASA descriptors which was then used to score the remaining ligands in the screening databases. The performance of SASA descriptors was found comparable or superior to those of GOLD and Glide.
Here, we compared the results of a fast Surflex docking-scoring protocol against rescoring with five GOLD scoring functions. In addition, we evaluated the screening potential of docking-pose derived protein-, ligand-, and protein–ligand-SASA descriptors as well as that of ML models developed with these descriptors while monitoring the impact on (early) enrichment factors. Similar to Surflex, GOLD scoring functions were also found to be target-dependent. Among all GOLD scoring functions, GoldScore produced the best AUC value in one protein structure of Mcl-1 (5MES) and both structures of Mdm2. ASP generated the best results in Bcl-2 (2O21), BRD4-1, and CREBBP (5EIC). The performance of other GOLD scoring functions was lower than the one of Surflex. In contrast to traditional scoring functions, the overall performance and early database enrichments for many SASA descriptors were found superior to Surflex or GOLD scoring functions for a large variety of target classes suggesting that these scoring metrics could be interesting for structure-based screening experiments. For instance, among the protein SASA descriptors, SASA receptor total delta ΔTR was the best-performing descriptor that produced excellent AUC values in nine protein structures. This indicates that the difference in the unbound and the bound SASA area of the binding site is a unique metric for each ligand that could be used to discriminate actives from inactive molecules. Among the ligand SASA descriptors, SASA ligand total unbound TLU produced the best results in eight target structures. This is consistent with previous studies which suggested that the rSASA score of the ligand, which is derived from the bound and the unbound fraction of the ligand SASA, can be used for finding ligands that are largely solvent-exposed at the PPI sites [41,50]. The rSASA of the active PPI ligand is relatively lower as compared to that of the inactive compound. Among the protein–ligand SASA descriptor, SASA receptor–ligand total delta ΔTRL generated excellent results in eight target structures. This suggests that the net SASA of the protein–ligand complex is another distinctive feature of an active ligand that can be valuable for screening purposes. The vs. performance was further enhanced when we used the descriptors for training ML models and validated them on test sets. Among the ML models and the present PPI dataset, neural networks and random forest achieved the best results in most of the targets except BAZ2B, for which all methods could not yield desirable results. In the case of Menin, the results were only satisfactory using random forest models. The visual inspection of the binding sites of BAZ2B structures showed relatively tight subpockets that are formed by hydrophobic/aromatic residues and with a conserved water molecule. In the case of Menin structures, the binding site is large and substantially flatter compared to the other PPIs. These features pose a significant challenge in ligand docking and it is possible that the poses generated for the compounds on these targets are not accurate enough. This in turn leads to the calculation of incorrect SASA descriptors that could not help train good ML models. Importantly, a consistent trend is observed where all methods (native docking, rescoring, SASA metrics, and ML models) failed on these targets which likely points to the fact that docking poses are unreliable and/or that some possible experimental errors could be present.
As more information is being used when developing ML models as compared to only relying on a docking score, as expected, the results obtained from the ML approaches are superior to the best-rescoring function DLIGAND2 identified in our last study [50] both in terms of AUC and early enrichments. Additionally, we would like to emphasize that in the future we would be working in a more flexible ML platform that could allow automatic tuning of hyperparameters. Since we are dealing with large datasets, different targets, and ML techniques, it was computationally expensive to manually explore all hyperparameters in Pipeline Pilot. However, it became clear from our experiments that adopting default hyperparameters showed a consistently good performance in most of the targets. Yet we anticipate that the performances observed could be further improved (particularly sensitivity and early enrichments) after tuning hyperparameters.
In trying to rationalize why SASA descriptors would help in discriminating between good and bad binders, we already mentioned that the method does not seem to favor high MW molecules [50]. In fact, it would seem that the SASA descriptors as used here assist the selection of molecules that fill better the binding site most likely important for PPIs as it has been shown that ligand binding pockets within a PPI interface are very different from traditional ligand binding pockets and tend to contain three to five binding sub-pockets of low volume [56,57]. It is also possible that SASA descriptors are fuzzier than interaction fingerprints or scoring functions and as such less sensitive to possible docking errors likely to occur due to the plasticity of most PPI interfaces. Thus, they can be interesting to combine with such approaches while important work will be required to fully investigate this question. For the time being, combining SASA descriptors with rescoring is still target-dependent, but the present results suggest that the approach could be interesting to circumvent in part this problem.
The protocol presented here, which involves docking compounds, extracting 3D descriptors from the docking poses, and using them for building ML models, seems interesting to investigate further with regard to the generation of compound collections potentially enriched in PPI modulators. We suggest that pose-derived SASA descriptors or pose-derived SASA descriptors combined with ML techniques could be a useful adjunct to other methods. For example, most of the work done so far to design dedicated PPI collections are mainly ligand-based. For instance, Hamon et al. built an SVM model with DRAGON molecular descriptors for small molecules using the 2P2I database (https://2p2idb.marseille.inserm.fr, accessed on 14 November 2022) to define a physicochemical profile of orthosteric inhibitors. The model was successfully used to mine PPI-like compounds from external datasets from PubChem BioAssay and in-house small molecule collections [47]. In another study, Reynes et al., by constructing decision trees on a dataset of known PPI inhibitors and regular drugs, determined a global physicochemical profile for the putative PPI inhibitors. The statistical analysis revealed two important molecular descriptors for PPI inhibitors characterizing specific molecular shapes and the presence of a privileged number of aromatic bonds. The best model was transposed into a computer program, PPI-HitProfiler, that can output from any drug-like compound collection a focused chemical library enriched in putative PPI inhibitors [58]. In a recent study, Bosc et al. constructed different ML models based on molecular descriptors using databases of experimentally confirmed PPI inhibitors. The predictive models were used to curate the putative PPI inhibitors from large commercial compound collections. The curated collection is available on-demand to the scientific community in 384-well plates as the Fr-PPIChem library. The compounds from the library showed a 46-fold activity rate enhancement compared to a non-iPPI-enriched diversity library in high-throughput screening against the CD47-SIRPα PPI [59]. It seems possible that SASA descriptors generated from docking poses could also be valuable to assist the design of PPI collections.
In summary, we showcased a proof-of-concept study highlighting the potential of GOLD scoring functions; several post-docking derivatized SASA descriptors of protein, ligand, and protein–ligand complex; and ML models to enhance the performance of fast docking computations performed with Surflex (“pscreen” mode) as monitored by ROC AUC and early database enrichment efficiency using experimentally known inhibitors of PPI targets. Significantly higher AUC values and enrichment factors were obtained for several targets using 3D SASA descriptors combined with rescoring or combined with ML as compared to default “pscreen” Surflex computations or rescoring using only GOLD functions, indicating an improvement in both the global ranking of compounds and also, in the early enrichment stages. The latter significant issue is additionally manifested in ligand enrichments at 1%, which were augmented by up to seven times in several proteins using neural network models. Collectively, our results strongly encourage the use of docking in combination with the use of pose-derived SASA descriptors or with ML techniques when a sufficient number of bioactivity data are available, to screen large molecular databases in search of new PPI inhibitors.

4. Materials and Methods

4.1. Rescoring with GOLD Scoring Functions and Consensus Approach

The procedure of docking PPI datasets to the respective protein structures is described in ref [50]. The docking poses obtained from each protein structure were rescored using five different scoring functions with default settings: (i) GoldScore: It is the original scoring function provided with GOLD [51,60]. It has been optimized for the prediction of ligand binding positions and takes into account factors such as hydrogen bonding energy, van der Waals energy, and ligand torsion strain; (ii) ChemPLP: piecewise linear potential is an empirical fitness function optimized for pose prediction. The piecewise linear potential (PLP) is used to model the steric complementarity between protein and ligand, but additionally, in ChemPLP, the distance- and angle-dependent hydrogen and metal bonding terms from Chemscore are considered. ChemPLP is slightly faster than ChemScore and up to four times faster than GoldScore. Several validation tests have shown it to be generally more effective than the other scoring functions for both pose prediction and virtual screening; (iii) Astex Statistical Potential (ASP): ASP is an atom-atom potential derived from a database of protein–ligand complexes and can be compared to other knowledge-based scoring potentials, e.g., PMF and DrugScore. ASP has comparable accuracy to ChemScore and GoldScore; (iv) ChemScore: It estimates the total free energy change that occurs on ligand binding and was trained by regression against binding affinity data for 82 complexes. The ChemScore fitness function also incorporates a protein–ligand atom clash term and an internal energy term. ChemScore takes into account hydrophobic-hydrophobic contact area, hydrogen bonding, ligand flexibility, and metal interaction; and (v) ChemScore (RDS), a modified form of ChemScore where the effect of scaling hydrogen bonds, metal–ligand, and hydrophobic interactions are systematically investigated based on the burriedness of intermolecular interactions in protein–ligand complexes. Burriedness is expressed using receptor density scaling (RDS), as defined by the number of heavy atoms around an interaction. Optimization of scaled ChemScore terms using burriedness from experimental X-ray data and docked inactive ligands has shown to improve the ability to discriminate actives from inactives significantly. These modified ChemScore terms penalize interactions with flexible outer walls or loops of the binding site; they adjust hydrogen-bond strengths in solvent-exposed surface areas when there is competition with solvent for favorable hydrogen bond interactions [51].
Consensus scoring is a method in which the predicted binding affinities or scores of each compound for a binding pocket are predicted by using more than one scoring method. In this study, a consensus scoring approach using the ‘rank-by-rank’ method [33] was applied to evaluate the hits obtained from high throughput docking. All the candidates were ranked by the average ranks predicted by all the scoring functions. This strategy uses relative ranks rather than absolute binding affinities for ranking. Since the compound docking scores obtained from Surflex and GOLD rescoring are of different nature and type, data normalization was performed to bring all the scores to a common scale ranging from 0 to 1. Data normalization was performed using Equations (1) and (2).
N o r m a l i z e d   s c o r e = D o c k i n g   s c o r e D o c k i n g   s c o r e m i n D o c k i n g   s c o r e m a x D o c k i n g   s c o r e m i n  
The compounds were then ranked based on the normalized docking scores from the five scoring functions. Finally, the results from different scoring functions were combined by averaging the rank of each molecule obtained from the individual scoring function (Equation (2)). These consensus scores were then utilized for rescoring and vs. assessment. The compounds were ranked from best to worst based on their consensus rank.
C o n s e n s u s   R a n k = S u r f l e x r a n k + G o l d S c o r e r a n k + P L P r a n k + A S P r a n k   + C h e m S c o r e r a n k + C h e m S c o r e R D S r a n k 6  

4.2. Solvent Accessible Surface Area (SASA) Calculations and Derivatization of SASA Descriptors

The SASA of the ligand and corresponding receptor-binding site in their bound/unbound states was calculated for each docking pose using the “binding_sasa.py” script from Schrödinger. More details on the procedure of calculation of SASA descriptors are described in ref [50]. Briefly, seven SASA descriptor subtypes (Hydrophobic (H), Aromatic (Ar), Donor (D), Acceptor (Ac), Positive (P), Negative (N), and Others (O)) were calculated for the protein and the ligand in the bound and unbound state. The protein and ligand descriptors were combined to generate the corresponding protein–ligand SASA descriptors. The total delta and relative total SASA for the receptor, ligand, and receptor–ligand were calculated as described in Equations (3)–(8). The relative SASA for each SASA category was calculated for ligand, protein, and protein–ligand using Equations (9)–(11). To evaluate the vs. performance of rescoring with GOLD scoring functions, consensus method, and SASA descriptors, we followed the protocol defined in Figure S1 for each PPI target. The performance was assessed by computing ROC graphs and calculating the AUC, EF1%, EF5%, and BEDROC (α = 20), demonstrating the ability of the vs. method to distinguish known ligands among the set of inactive compounds or predominance of active ligands in the top positions of the ranked list [61,62,63,64,65].
S A S A   t o t a l   l i g a n d   d e l t a   ( T L ) = S A S A   t o t a l   l i g a n d     u n b o u n d   ( T L U ) S A S A   t o t a l   l i g a n d   b o u n d   ( T L B )    
S A S A   t o t a l   r e c e p t o r   d e l t a   ( Δ T R ) = S A S A   t o t a l   r e c e p t o r   u n b o u n d   ( T R U ) S A S A   t o t a l   r e c e p t o r   b o u n d   ( T R B )    
S A S A   t o t a l   r e c e p t o r l i g a n d   d e l t a   ( Δ T R L ) = { S A S A   t o t a l   r e c e p t o r   u n b o u n d   ( T R U ) + S A S A   t o t a l   l i g a n d     u n b o u n d   ( T L U ) } { S A S A   t o t a l   r e c e p t o r   b o u n d   ( T R B ) + S A S A   t o t a l   r e c e p t o r   b o u n d   ( T L B ) }
r e l a t i v e   S A S A   T o t a l   l i g a n d   ( r e l .   T L ) = T L B T L U    
r e l a t i v e   S A S A   T o t a l   r e c e p t o r   ( r e l .   T R ) = T R B T R U
r e l a t i v e   S A S A   t o t a l   r e c e p t o r     l i g a n d   d e l t a   ( r e l .     T R L ) = T R B + T L B   T R U + T L U      
r e l a t i v e   S A S A   X   l i g a n d   ( r e l .   X L ) = X L B X L U         { W h e r e   X : H y d r o p h o b i c   ( H ) ,   A r o m a t i c   ( A r ) ,   D o n o r   ( D ) ,   A c c e p t o r   ( A c ) ,   P o s i t i v e   ( P ) ,   N e g a t i v e   ( N ) a n d   O t h e r s   ( O ) }
r e l a t i v e   S A S A   X   R e c e p t o r   ( r e l .   X R ) = X R B X R U         { W h e r e   X : H y d r o p h o b i c   ( H ) ,   A r o m a t i c   ( A r ) ,   D o n o r   ( D ) ,   A c c e p t o r   ( A c ) ,   P o s i t i v e   ( P ) ,   N e g a t i v e   ( N ) a n d   O t h e r s   ( O ) }
r e l a t i v e   S A S A X   r e c e p t o r   l i g a n d   d e l t a ( Δ T R L ) = X R B + X L B X R U + X L U { W h e r e   X : H y d r o p h o b i c   ( H ) ,   A r o m a t i c   ( A r ) ,   D o n o r   ( D ) ,   A c c e p t o r   ( A c ) ,   P o s i t i v e   ( P ) ,   N e g a t i v e   ( N )   a n d   O t h e r s   ( O ) }

4.3. Machine Learning Models Generation

All ML models were generated using the BIOVIA pipeline pilot v18.1 interfaced with the R package V 3.4.1 [66,67] using the recommended standard protocol. All models were developed using the same set of SASA descriptors. The SASA scores were normalized before using them for building ML models. To address the dataset imbalance, the weighting method was set to “by class” during the model’s construction. Three types of Recursive Partitioning (RP) models were constructed: tree, bagged forest, and random forest. For creating RP single decision tree models, following parameters were used: model type = single tree; Minimum samples per node = 10; Maximum tree depth = 20; Split method = Gini; Weighting Method = By Class; Maximum Knots per property = 20; Maximum look ahead depth = 0; Maximum generic depth = 0. RP Bagged forest models are an extension of the decision tree approach that minimizes errors from over-fitting by generating multiple trees and randomizing the training sets (without resampling the descriptors) that each tree utilizes to reproduce the trends in the training set. “Bagging” (bootstrap aggregation) with replacement is used to generate a new, modified version of the training set for each tree: each tree tends to get a random subset of the original training set, with ~2/3 unique compounds and ~1/3 duplicates. In the case of RP bagged forest models, ten trees (model type = Forest) were created with “bagging” with the same settings as mentioned above. In the case of the forest of random trees (random forest), 500 trees were created with “bagging” with a number of randomly preselected variables set to 9 to be considered as splitting criteria for each node. In a Random forest, when each tree is “induced” (created), it receives a different subset of the descriptors and the algorithm applies different weights on the descriptors, which can also be used in different orders in the different hierarchies of the levels of nodes within each tree. Each tree then develops a different model to reproduce its training set. Each tree’s classification of the test instance is recorded as a vote. The votes from all trees are aggregated and the test instance is assigned to the class that receives the maximum vote. We also generated SVM-based models which are based on the principle of finding a plane that best separates a description of the two classes of compounds in the training set. For complicated datasets, algorithms called kernel functions are used to transform the descriptors of a dataset into a higher dimensional space, to better separate the “good” and “bad” compounds. Support vectors are constructed that use a minimal number of the “good” and “bad” compounds to define a boundary to the hyper plane (the multi-dimensional plane that separates the transformed descriptions of the compounds in higher dimensional space), in a way that attempts to minimize error while maximizing generalizability. The parameters used were as follows: SVM-Type = C-classification; SVM-Kernel = radial; Cost = 1, 2; Weighting Method = By Class; Gamma = 0.0125. The logistic regression models were constructed using the GLM method (Generalized linear model). GLM performs a logistic regression with no bias correction. The deep neural net models were constructed with two hidden layers and each layer with 80 nodes. The learning rate of every epoch was 0.1 with a momentum of 0.9, and the maximum number of iterations for network training was 5000. To prevent the model from over-fitting, the fraction of the hidden layer to be dropped out for model training was set to 0.25. Bayesian models build a binary classifier by uniting different sets of descriptors, with different weights on each descriptor, in diverse ways, to build a model that best reproduces the known trends in the training set. The Bayesian models were constructed using the naïve Bayes algorithm. Naïve Bayes is a fast, scalable algorithm that calculates conditional probabilities for combinations of attributes and the target attribute. From the training data, an independent probability is established. This probability gives the likelihood of each target class, given the occurrence of each value category from each input variable. It is known to perform well on large datasets and has very fast processing times. In all cases, 10-fold cross-validation was used to calculate the ROC curve. In addition, we calculated several metrics, such as Sensitivity (Equation (12)), specificity (Equation (13)), precision (Equation (14)), and concordance (Equation (15)), to evaluate the performances of classifiers. Sensitivity represents the percentage of correctly identified active compounds. Specificity signifies the percentage of correctly identified inactive compounds. Concordance corresponds to the overall accuracy. Precision signifies the percentage of identified positive compounds or positive predictive value (PPV). In addition, we calculated Youden’s J statistic (Equation (16)), Matthew’s correlation coefficient (MCC) (Equation (17)), the F1 score (Equation (18)), and Cohen’s Kappa statistic (Cκ or κ) (Equation (19)). Youden’s index (J) is a performance metric that evaluates the performance of a binary classification model. Its value ranges from 0 to 1. When the value is at its minimum (i.e., zero), the model is useless. When its value is 1, there are no false negatives or false positives, and the predictions are perfect. MCC, also referred to as the phi coefficient, is a chance-corrected statistic where MCC = 1 indicates perfect agreement, MCC = −1 indicates total disagreement, and MCC = 0 indicates that the model is no better than random. The F1 score is the harmonic mean of precision (i.e., the positive predictive value hit rate) and sensitivity. The best F1 score is 1, while the worst score possible is 0. Cohen’s Kappa (κ) is a chance-corrected statistic that uses a different method to calculate the random likelihood of making correct predictions for the external set. Kappa < 0 indicates no agreement, Kappa of 0–0.2 indicates slight agreement, Kappa of 0.21–0.40 is fairly predictive, and Kappa of 0.41–0.60 indicates moderate agreement.
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P
P r e c i s i o n = T P T P + F P
C o n c o r d a n c e   ( A c c u r a c y ) = T P + T N T P + T N + F P + T N
J = s e n s i t i v i t y + s p e c i f i c i t y 1
M C C = ( T P × T N ) ( F P × F N ) ( T P + F P ) × ( T P + F N ) × ( T N + F P ) × ( T N + F N )
F 1 = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l
κ = P o P E 1 P E = 1 1 P o 1 P E
In Equation (15), true positive (TP) is the number of active ligands that are predicted correctly; true negative (TN) is the number of inactive ligands that are predicted correctly; false negative (FN) is the number of active ligands that are predicted as inactive molecules, and false positive (FP) is the number of inactive ligands that are predicted to be active. In Equation (19), PO is the relative observed agreement among raters, and PE is the hypothetical probability of chance agreement, obtained using the observed data to calculate the probabilities of each observer randomly saying each category. If the raters are in complete agreement, then κ = 1. If there is no agreement among the raters other than what would be expected by chance (as given by PE), κ ≤ 0.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ijms232214364/s1.

Author Contributions

N.S. and B.O.V. designed the study; N.S. performed the computations; N.S. analyzed the results; N.S. provided a first draft of the paper, and B.O.V. reviewed the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

We thank the INSERM institute for financial support.

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

PPIs, protein-protein interactions; VS, virtual screening; SBVS, structure-based virtual screening; PDB, protein data bank; ASP, astex statistical potential; RDS, receptor depth scaling; ML, machine learning; SVM, support vector machine, QSAR, quantitative structure-activity relationship; rSASA, relative solvent accessible surface area; ROC, receiver operating characteristic; AUC, area under the ROC curve; BEDROC, Boltzmann-enhanced discrimination of ROC; EF 1%, enrichment factor at 1%; EF 5%, enrichment factor at 5%; BAZ2B, Bromodomain Adjacent to Zinc Finger Domain 2B; BRD4-1, Bromodomain-containing protein 4; CREBBP, CREB-binding protein; HIV IN, HIV Integrase; XIAP, Inhibitor of apoptosis protein 3; EphA4, Ephrin type-A receptor 4; PCA, principal component analysis; PLIF, protein–ligand interaction fingerprint (PLIF).

References

  1. Walters, W.P.; Stahl, M.T.; Murcko, M.A. Virtual Screening—An Overview. Drug Discov. Today 1998, 3, 160–178. [Google Scholar] [CrossRef]
  2. Shoichet, B.K. Virtual Screening of Chemical Libraries. Nature 2004, 432, 862–865. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Oprea, T.I.; Matter, H. Integrating Virtual Screening in Lead Discovery. Curr. Opin. Chem. Biol. 2004, 8, 349–358. [Google Scholar] [CrossRef] [PubMed]
  4. Bajorath, J. Integration of Virtual and High-Throughput Screening. Nat. Rev. Drug Discov. 2002, 1, 882–894. [Google Scholar] [CrossRef]
  5. Bissantz, C.; Folkers, G.; Rognan, D. Protein-Based Virtual Screening of Chemical Databases. 1. Evaluation of Different Docking/Scoring Combinations. J. Med. Chem. 2000, 43, 4759–4767. [Google Scholar] [CrossRef]
  6. Ma, D.-L.; Chan, D.S.-H.; Leung, C.-H. Drug Repositioning by Structure-Based Virtual Screening. Chem. Soc. Rev. 2013, 42, 2130–2141. [Google Scholar] [CrossRef]
  7. Lyne, P.D. Structure-Based Virtual Screening: An Overview. Drug Discov. Today 2002, 7, 1047–1055. [Google Scholar] [CrossRef]
  8. Mirdita, M.; Schütze, K.; Moriwaki, Y.; Heo, L.; Ovchinnikov, S.; Steinegger, M. ColabFold: Making Protein Folding Accessible to All. Nat. Methods 2022, 19, 679–682. [Google Scholar] [CrossRef]
  9. Jumper, J.; Evans, R.; Pritzel, A.; Green, T.; Figurnov, M.; Ronneberger, O.; Tunyasuvunakool, K.; Bates, R.; Žídek, A.; Potapenko, A.; et al. Highly Accurate Protein Structure Prediction with AlphaFold. Nature 2021, 596, 583–589. [Google Scholar] [CrossRef]
  10. Baek, M.; DiMaio, F.; Anishchenko, I.; Dauparas, J.; Ovchinnikov, S.; Lee, G.R.; Wang, J.; Cong, Q.; Kinch, L.N.; Schaeffer, R.D.; et al. Accurate Prediction of Protein Structures and Interactions Using a Three-Track Neural Network. Science 2021, 373, 871–876. [Google Scholar] [CrossRef]
  11. Bryant, P.; Pozzati, G.; Elofsson, A. Improved Prediction of Protein-Protein Interactions Using AlphaFold2. Nat. Commun. 2022, 13, 1265. [Google Scholar] [CrossRef]
  12. Wigge, C.; Stefanovic, A.; Radjainia, M. The Rapidly Evolving Role of Cryo-EM in Drug Design. Drug Discov. Today Technol. 2020, 38, 91–102. [Google Scholar] [CrossRef]
  13. Van Drie, J.H.; Tong, L. Cryo-EM as a Powerful Tool for Drug Discovery. Bioorg. Med. Chem. Lett. 2020, 30, 127524. [Google Scholar] [CrossRef] [PubMed]
  14. Ceska, T.; Chung, C.-W.; Cooke, R.; Phillips, C.; Williams, P.A. Cryo-EM in Drug Discovery. Biochem. Soc. Trans. 2019, 47, 281–293. [Google Scholar] [CrossRef] [Green Version]
  15. Rognan, D. The Impact of in Silico Screening in the Discovery of Novel and Safer Drug Candidates. Pharmacol. Ther. 2017, 175, 47–66. [Google Scholar] [CrossRef]
  16. Slater, O.; Kontoyianni, M. The Compromise of Virtual Screening and Its Impact on Drug Discovery. Expert Opin. Drug Discov. 2019, 14, 619–637. [Google Scholar] [CrossRef] [PubMed]
  17. Gimeno, A.; Ojeda-Montes, M.J.; Tomás-Hernández, S.; Cereto-Massagué, A.; Beltrán-Debón, R.; Mulero, M.; Pujadas, G.; Garcia-Vallvé, S. The Light and Dark Sides of Virtual Screening: What Is There to Know? Int. J. Mol. Sci. 2019, 20, 1375. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Spyrakis, F.; Cavasotto, C.N. Open Challenges in Structure-Based Virtual Screening: Receptor Modeling, Target Flexibility Consideration and Active Site Water Molecules Description. Arch. Biochem. Biophys. 2015, 583, 105–119. [Google Scholar] [CrossRef]
  19. Scior, T.; Bender, A.; Tresadern, G.; Medina-Franco, J.L.; Martínez-Mayorga, K.; Langer, T.; Cuanalo-Contreras, K.; Agrafiotis, D.K. Recognizing Pitfalls in Virtual Screening: A Critical Review. J. Chem. Inf. Model 2012, 52, 867–881. [Google Scholar] [CrossRef]
  20. Plewczynski, D.; Łaźniewski, M.; Augustyniak, R.; Ginalski, K. Can We Trust Docking Results? Evaluation of Seven Commonly Used Programs on PDBbind Database. J. Comput. Chem. 2011, 32, 742–755. [Google Scholar] [CrossRef]
  21. Sheridan, R.P.; Kearsley, S.K. Why Do We Need so Many Chemical Similarity Search Methods? Drug Discov. Today 2002, 7, 903–911. [Google Scholar] [CrossRef]
  22. Yang, J.; Shen, C.; Huang, N. Predicting or Pretending: Artificial Intelligence for Protein-Ligand Interactions Lack of Sufficiently Large and Unbiased Datasets. Front. Pharmacol. 2020, 11, 69. [Google Scholar] [CrossRef] [PubMed]
  23. Shen, C.; Hu, Y.; Wang, Z.; Zhang, X.; Zhong, H.; Wang, G.; Yao, X.; Xu, L.; Cao, D.; Hou, T. Can Machine Learning Consistently Improve the Scoring Power of Classical Scoring Functions? Insights into the Role of Machine Learning in Scoring Functions. Brief. Bioinform. 2020, 22, 497–514. [Google Scholar] [CrossRef] [PubMed]
  24. Boyles, F.; Deane, C.M.; Morris, G.M. Learning from the Ligand: Using Ligand-Based Features to Improve Binding Affinity Prediction. Bioinformatics 2020, 36, 758–764. [Google Scholar] [CrossRef]
  25. Torres, P.H.M.; Sodero, A.C.R.; Jofily, P.; Silva-Jr, F.P. Key Topics in Molecular Docking for Drug Design. Int. J. Mol. Sci. 2019, 20, 4574. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Li, J.; Fu, A.; Zhang, L. An Overview of Scoring Functions Used for Protein-Ligand Interactions in Molecular Docking. Interdiscip. Sci. 2019, 11, 320–328. [Google Scholar] [CrossRef]
  27. Sieg, J.; Flachsenberg, F.; Rarey, M. In Need of Bias Control: Evaluating Chemical Data for Machine Learning in Structure-Based Virtual Screening. J. Chem. Inf. Model. 2019, 59, 947–961. [Google Scholar] [CrossRef]
  28. Nogueira, M.S.; Koch, O. The Development of Target-Specific Machine Learning Models as Scoring Functions for Docking-Based Target Prediction. J. Chem. Inf. Model. 2019, 59, 1238–1252. [Google Scholar] [CrossRef]
  29. Guedes, I.A.; Pereira, F.S.S.; Dardenne, L.E. Empirical Scoring Functions for Structure-Based Virtual Screening: Applications, Critical Aspects, and Challenges. Front. Pharmacol. 2018, 9, 1089. [Google Scholar] [CrossRef]
  30. Wingert, B.M.; Camacho, C.J. Improving Small Molecule Virtual Screening Strategies for the next Generation of Therapeutics. Curr. Opin. Chem. Biol. 2018, 44, 87–92. [Google Scholar] [CrossRef]
  31. Wójcikowski, M.; Ballester, P.J.; Siedlecki, P. Performance of Machine-Learning Scoring Functions in Structure-Based Virtual Screening. Sci. Rep. 2017, 7, 46710. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Feher, M. Consensus Scoring for Protein-Ligand Interactions. Drug Discov. Today 2006, 11, 421–428. [Google Scholar] [CrossRef] [PubMed]
  33. Wang, R.; Wang, S. How Does Consensus Scoring Work for Virtual Library Screening? An Idealized Computer Experiment. J. Chem. Inf. Comput. Sci. 2001, 41, 1422–1426. [Google Scholar] [CrossRef]
  34. Singh, N.; Villoutreix, B.O. Demystifying the Molecular Basis of Pyrazoloquinolinones Recognition at the Extracellular A1+/Β3- Interface of the GABAA Receptor by Molecular Modeling. Front. Pharmacol. 2020, 11, 561834. [Google Scholar] [CrossRef] [PubMed]
  35. Arkin, M.R.; Wells, J.A. Small-Molecule Inhibitors of Protein-Protein Interactions: Progressing towards the Dream. Nat. Rev. Drug Discov. 2004, 3, 301–317. [Google Scholar] [CrossRef]
  36. Stumpf, M.P.H.; Thorne, T.; de Silva, E.; Stewart, R.; An, H.J.; Lappe, M.; Wiuf, C. Estimating the Size of the Human Interactome. Proc. Natl. Acad. Sci. USA 2008, 105, 6959–6964. [Google Scholar] [CrossRef] [Green Version]
  37. Venkatesan, K.; Rual, J.-F.; Vazquez, A.; Stelzl, U.; Lemmens, I.; Hirozane-Kishikawa, T.; Hao, T.; Zenkner, M.; Xin, X.; Goh, K.-I.; et al. An Empirical Framework for Binary Interactome Mapping. Nat. Methods 2009, 6, 83–90. [Google Scholar] [CrossRef]
  38. Cheng, A.C.; Coleman, R.G.; Smyth, K.T.; Cao, Q.; Soulard, P.; Caffrey, D.R.; Salzberg, A.C.; Huang, E.S. Structure-Based Maximal Affinity Model Predicts Small-Molecule Druggability. Nat. Biotechnol. 2007, 25, 71–75. [Google Scholar] [CrossRef]
  39. Blundell, T.L.; Burke, D.F.; Chirgadze, D.; Dhanaraj, V.; Hyvönen, M.; Innis, C.A.; Parisini, E.; Pellegrini, L.; Sayed, M.; Sibanda, B.L. Protein-Protein Interactions in Receptor Activation and Intracellular Signalling. Biol. Chem. 2000, 381, 955–959. [Google Scholar] [CrossRef]
  40. Chen, P.; Ke, Y.; Lu, Y.; Du, Y.; Li, J.; Yan, H.; Zhao, H.; Zhou, Y.; Yang, Y. DLIGAND2: An Improved Knowledge-Based Energy Function for Protein-Ligand Interactions Using the Distance-Scaled, Finite, Ideal-Gas Reference State. J. Cheminform. 2019, 11, 52. [Google Scholar] [CrossRef]
  41. Trisciuzzi, D.; Nicolotti, O.; Miteva, M.A.; Villoutreix, B.O. Analysis of Solvent-Exposed and Buried Co-Crystallized Ligands: A Case Study to Support the Design of Novel Protein–Protein Interaction Inhibitors. Drug Discov. Today 2019, 24, 551–559. [Google Scholar] [CrossRef]
  42. Díaz-Eufracio, B.I.; Medina-Franco, J.L. Towards the Development of Machine Learning Models to Predict Protein-Protein Interaction Modulators. ChemRxiv 2022. [Google Scholar] [CrossRef]
  43. Sarkar, D.; Saha, S. Machine-Learning Techniques for the Prediction of Protein–Protein Interactions. J. Biosci. 2019, 44, 104. [Google Scholar] [CrossRef] [PubMed]
  44. Gupta, P.; Mohanty, D. SMMPPI: A Machine Learning-Based Approach for Prediction of Modulators of Protein-Protein Interactions and Its Application for Identification of Novel Inhibitors for RBD:HACE2 Interactions in SARS-CoV-2. Brief Bioinform. 2021, 22, bbab111. [Google Scholar] [CrossRef]
  45. Neugebauer, A.; Hartmann, R.W.; Klein, C.D. Prediction of Protein-Protein Interaction Inhibitors by Chemoinformatics and Machine Learning Methods. J. Med. Chem. 2007, 50, 4665–4668. [Google Scholar] [CrossRef] [PubMed]
  46. Sperandio, O.; Reynès, C.H.; Camproux, A.-C.; Villoutreix, B.O. Rationalizing the Chemical Space of Protein–Protein Interaction Inhibitors. Drug Discov. Today 2010, 15, 220–229. [Google Scholar] [CrossRef]
  47. Hamon, V.; Bourgeas, R.; Ducrot, P.; Theret, I.; Xuereb, L.; Basse, M.J.; Brunel, J.M.; Combes, S.; Morelli, X.; Roche, P. 2P2I HUNTER: A Tool for Filtering Orthosteric Protein-Protein Interaction Modulators via a Dedicated Support Vector Machine. J. R. Soc. Interface 2014, 11, 20130860. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Gaulton, A.; Bellis, L.J.; Bento, A.P.; Chambers, J.; Davies, M.; Hersey, A.; Light, Y.; McGlinchey, S.; Michalovich, D.; Al-Lazikani, B.; et al. ChEMBL: A Large-Scale Bioactivity Database for Drug Discovery. Nucleic. Acids Res. 2012, 40, D1100–D1107. [Google Scholar] [CrossRef] [Green Version]
  49. Kim, S.; Thiessen, P.A.; Bolton, E.E.; Chen, J.; Fu, G.; Gindulyte, A.; Han, L.; He, J.; He, S.; Shoemaker, B.A.; et al. PubChem Substance and Compound Databases. Nucleic. Acids Res. 2016, 44, D1202–D1213. [Google Scholar] [CrossRef]
  50. Singh, N.; Chaput, L.; Villoutreix, B.O. Fast Rescoring Protocols to Improve the Performance of Structure-Based Virtual Screening Performed on Protein-Protein Interfaces. J. Chem. Inf. Model. 2020, 60, 3910–3934. [Google Scholar] [CrossRef]
  51. Verdonk, M.L.; Cole, J.C.; Hartshorn, M.J.; Murray, C.W.; Taylor, R.D. Improved Protein-Ligand Docking Using GOLD. Proteins 2003, 52, 609–623. [Google Scholar] [CrossRef]
  52. Whitty, A.; Kumaravel, G. Between a Rock and a Hard Place? Nat. Chem. Biol. 2006, 2, 112–118. [Google Scholar] [CrossRef]
  53. Fry, D.C. Drug-like Inhibitors of Protein-Protein Interactions: A Structural Examination of Effective Protein Mimicry. Curr. Protein Pept. Sci. 2008, 9, 240–247. [Google Scholar] [CrossRef]
  54. Chène, P. Drugs Targeting Protein-Protein Interactions. ChemMedChem 2006, 1, 400–411. [Google Scholar] [CrossRef] [PubMed]
  55. Núñez, S.; Venhorst, J.; Kruse, C.G. Assessment of a Novel Scoring Method Based on Solvent Accessible Surface Area Descriptors. J. Chem. Inf. Model. 2010, 50, 480–486. [Google Scholar] [CrossRef]
  56. Kuenemann, M.A.; Sperandio, O.; Labbé, C.M.; Lagorce, D.; Miteva, M.A.; Villoutreix, B.O. In Silico Design of Low Molecular Weight Protein-Protein Interaction Inhibitors: Overall Concept and Recent Advances. Prog. Biophys. Mol. Biol. 2015, 119, 20–32. [Google Scholar] [CrossRef]
  57. Fuller, J.C.; Burgoyne, N.J.; Jackson, R.M. Predicting Druggable Binding Sites at the Protein-Protein Interface. Drug Discov. Today 2009, 14, 155–161. [Google Scholar] [CrossRef]
  58. Reynès, C.; Host, H.; Camproux, A.-C.; Laconde, G.; Leroux, F.; Mazars, A.; Deprez, B.; Fahraeus, R.; Villoutreix, B.O.; Sperandio, O. Designing Focused Chemical Libraries Enriched in Protein-Protein Interaction Inhibitors Using Machine-Learning Methods. PLoS Comput. Biol. 2010, 6, e1000695. [Google Scholar] [CrossRef] [PubMed]
  59. Bosc, N.; Muller, C.; Hoffer, L.; Lagorce, D.; Bourg, S.; Derviaux, C.; Gourdel, M.-E.; Rain, J.-C.; Miller, T.W.; Villoutreix, B.O.; et al. Fr-PPIChem: An Academic Compound Library Dedicated to Protein–Protein Interactions. ACS Chem. Biol. 2020. [Google Scholar] [CrossRef] [PubMed]
  60. Jones, G.; Willett, P.; Glen, R.C.; Leach, A.R.; Taylor, R. Development and Validation of a Genetic Algorithm for Flexible Docking. J. Mol. Biol. 1997, 267, 727–748. [Google Scholar] [CrossRef]
  61. Triballeau, N.; Acher, F.; Brabet, I.; Pin, J.-P.; Bertrand, H.-O. Virtual Screening Workflow Development Guided by the “Receiver Operating Characteristic” Curve Approach. Application to High-Throughput Docking on Metabotropic Glutamate Receptor Subtype 4. J. Med. Chem. 2005, 48, 2534–2547. [Google Scholar] [CrossRef] [PubMed]
  62. Mysinger, M.M.; Shoichet, B.K. Rapid Context-Dependent Ligand Desolvation in Molecular Docking. J. Chem. Inf. Model. 2010, 50, 1561–1573. [Google Scholar] [CrossRef] [PubMed]
  63. Truchon, J.-F.; Bayly, C.I. Evaluating Virtual Screening Methods: Good and Bad Metrics for the “Early Recognition” Problem. J. Chem. Inf. Model. 2007, 47, 488–508. [Google Scholar] [CrossRef] [PubMed]
  64. Venkatraman, V.; Pérez-Nueno, V.I.; Mavridis, L.; Ritchie, D.W. Comprehensive Comparison of Ligand-Based Virtual Screening Tools against the DUD Data Set Reveals Limitations of Current 3D Methods. J. Chem. Inf. Model. 2010, 50, 2079–2093. [Google Scholar] [CrossRef] [Green Version]
  65. Empereur-mot, C.; Guillemain, H.; Latouche, A.; Zagury, J.-F.; Viallon, V.; Montes, M. Predictiveness Curves in Virtual Screening. J. Cheminform. 2015, 7, 52. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Warr, W.A. Scientific Workflow Systems: Pipeline Pilot and KNIME. J. Comput. Aided Mol. Des. 2012, 26, 801–804. [Google Scholar] [CrossRef]
  67. Gentleman, R.; Hornik, K.; Leisch, F. R 1.5 and the Bioconductor 1.0 Releases. Comput. Stat. Data An. 2002, 39, 557–558. [Google Scholar]
Figure 1. Heat map plot showing the AUC values for the ten PPI targets and their protein structures (y-axis) obtained using 32 protein SASA descriptors, 32 ligand SASA descriptors, and 16 protein-ligand SASA descriptors. The calculations were performed for each target-specific set of active and inactive binders.
Figure 1. Heat map plot showing the AUC values for the ten PPI targets and their protein structures (y-axis) obtained using 32 protein SASA descriptors, 32 ligand SASA descriptors, and 16 protein-ligand SASA descriptors. The calculations were performed for each target-specific set of active and inactive binders.
Ijms 23 14364 g001
Table 1. Overview of the ten PPI datasets used in this study. The target names, PDB codes of target protein structures, ChEMBL ID of the targets, the number of compounds categorized as “active” (activity class: 1) and “inactive” (activity class: 0) in the curated datasets, and the number of active and inactive compounds in the training and the test set for each target (after docking of the curated datasets) used in the construction and validation of the classification ML models, are indicated.
Table 1. Overview of the ten PPI datasets used in this study. The target names, PDB codes of target protein structures, ChEMBL ID of the targets, the number of compounds categorized as “active” (activity class: 1) and “inactive” (activity class: 0) in the curated datasets, and the number of active and inactive compounds in the training and the test set for each target (after docking of the curated datasets) used in the construction and validation of the classification ML models, are indicated.
Target NamePDB IDTarget ChEMBL IDNactivesNinactivesTraining Set (Nactives)Training Set (Ninactives)Test Set (Nactives)Test Set (Ninactives)
Bromodomain Adjacent to Zinc Finger Domain 2B (BAZ2B)4XUACHEMBL1741220685249,457479434,598205514,827
5E73
Apoptosis regulator Bcl-2 2O21CHEMBL4860178849,082124034,34353114,718
4LVT
Apoptosis regulator Bcl-xL 3INQCHEMBL462597149,19067434,42028914,751
3WIZ
BRD4 bromodomain 1 (BRD4-1)5D3LCHEMBL1163125847981592687253294
5KU3
CREB-binding protein (CREBBP)5EICCHEMBL5747136048,78191034,42539013,896
5MMG
HIV Integrase (HIV IN)4CFDCHEMBL23665059051232610855261366
4CHO
Inhibitor of apoptosis protein 3 (XIAP)1TFTCHEMBL4198114549,35179334,52834014,798
5C3H
Induced myeloid leukemia cell differentiation protein Mcl-15FC4CHEMBL4361145549,11299534,14742614,635
5MES
E3 ubiquitin-protein ligase Mdm24ODFCHEMBL502322274351155930406681303
4ZFI
Menin5DB2CHEMBL209386170531,51049122,0202119437
6B41
Table 2. The AUC and BEDROC scores (α = 20) for the ten PPI targets (BAZ2B: 4XUA, 5E73; Bcl-2: 2O21, 4LVT; Bcl-xL: 3INQ, 3WIZ; BRD4-1: 5D3L, 5KU3; CREBBP: 5EIC, 5MMG; HIV IN: 4CFD, 4CHO; XIAP: 1TFT, 5C3H; Mcl-1: 5FC4, 5MES; Mdm2: 4ODF, 4ZFI; and Menin: 5DB2, 6B41) and their corresponding protein structures obtained after rescoring of Surflex generated poses using five GOLD scoring functions (GoldScore, ChemScore, ASP, ChemPLP, and ChemScore RDS) and via consensus scheme involving all scoring functions. The best AUC and BEDROC values corresponding to each target protein (across rows) are shown in bold.
Table 2. The AUC and BEDROC scores (α = 20) for the ten PPI targets (BAZ2B: 4XUA, 5E73; Bcl-2: 2O21, 4LVT; Bcl-xL: 3INQ, 3WIZ; BRD4-1: 5D3L, 5KU3; CREBBP: 5EIC, 5MMG; HIV IN: 4CFD, 4CHO; XIAP: 1TFT, 5C3H; Mcl-1: 5FC4, 5MES; Mdm2: 4ODF, 4ZFI; and Menin: 5DB2, 6B41) and their corresponding protein structures obtained after rescoring of Surflex generated poses using five GOLD scoring functions (GoldScore, ChemScore, ASP, ChemPLP, and ChemScore RDS) and via consensus scheme involving all scoring functions. The best AUC and BEDROC values corresponding to each target protein (across rows) are shown in bold.
TargetPDB IDSurflexGoldScoreChemScoreASPChemPLPChemScore RDSConsensus Scoring
AUC BEDROCAUC BEDROCAUC BEDROCAUC BEDROCAUC BEDROCAUC BEDROCAUC BEDROC
BAZ2B4XUA0.4910.1200.4890.1270.5050.1440.4960.1390.4890.1270.5030.1320.4950.133
5E730.4820.1200.4820.1180.5070.1400.4930.1320.4870.1230.5060.1390.4930.127
Bcl-22O210.8570.6410.8600.6400.8390.5880.8830.5950.8530.7110.5900.2090.7980.548
4LVT0.8860.7010.8770.7000.8640.6900.8870.6940.8690.7390.6650.3680.8860.722
Bcl-xL3INQ0.8440.5230.7990.4630.7530.4080.7860.4110.7760.5130.5550.1880.7700.435
3WIZ0.8230.5000.8210.5210.7920.5100.8160.4610.8230.6140.5410.1740.8050.573
BRD4-15D3L0.6960.6720.7780.7770.7250.8080.8010.8180.7930.8580.6710.8520.7990.849
5KU30.4260.2540.6520.6000.5310.5160.6780.6740.6320.5720.5030.5300.6250.643
CREBBP5EIC0.6040.1000.6270.1140.5980.1230.6750.1590.6430.1340.5460.0820.6640.155
5MMG0.6620.1600.6540.1340.5790.1130.6310.1380.6130.1190.5400.0840.6200.135
HIV IN4CFD0.6270.6040.5900.3560.4620.4010.6140.5650.5950.4760.4510.3350.5670.533
4CHO0.6310.6390.6170.4250.4730.3710.6200.5890.6140.5370.4550.3350.5710.521
XIAP1TFT0.8880.7670.8510.6360.8230.5820.8110.4930.8450.5720.7520.5350.8060.563
5C3H0.8920.7040.6760.3410.7260.4920.7650.5020.7780.6320.7390.5510.7350.626
Mcl-15FC40.5850.1700.6360.2280.6570.2570.6580.2120.6420.2430.5950.2230.6590.241
5MES0.590.1500.6300.2050.6020.1670.5960.1520.6050.1660.5240.1300.6050.168
Mdm24ODF0.640.8420.7120.8690.6330.7910.7060.8320.6810.8220.5310.6490.6610.833
4ZFI0.570.6000.6570.7440.5350.4420.5940.5390.5540.4000.4350.2420.5660.492
Menin5DB20.5480.1040.5340.1360.5250.0990.5330.1240.5360.1220.5270.0870.5320.129
6B410.5170.0650.5210.1040.4780.0660.5090.0920.5270.0930.4690.0590.4890.076
Table 3. 10-fold cross-validation ROC AUC values with their standard deviations (SD) attained for ten PPI targets obtained using different ML learning approaches. The best AUC values corresponding to each target protein (across rows) are shown in bold.
Table 3. 10-fold cross-validation ROC AUC values with their standard deviations (SD) attained for ten PPI targets obtained using different ML learning approaches. The best AUC values corresponding to each target protein (across rows) are shown in bold.
TargetPDB IDTreeBagged ForestRandom ForestBayesianSVMLogistic RegressionNeural NetNeural Net (Bagging)
BAZ2B4XUA0.543 ± 0.0130.500 ± 0.0000.601 ± 0.145 0.602 ± 0.0010.758 ± 0.0050.550 ± 0.0180.543 ± 0.0090.651 ± 0.024
5E730.534 ± 0.0100.500 ± 0.0000.631 ± 0.1320.602 ± 0.0010.772 ± 0.0020.535 ± 0.0160.542 ± 0.0040.671 ± 0.021
Bcl-22O210.965 ± 0.011 0.943 ± 0.0010.999 ± 0.1170.984 ± 0.0000.990 ± 0.0000.973 ± 0.0000.992 ± 0.0010.999 ± 0.000
4LVT0.948 ± 0.0040.940 ± 0.0010.998 ± 0.0380.984 ± 0.0000.988 ± 0.0000.979 ± 0.0000.989 ± 0.0010.999 ± 0.000
Bcl-xL3INQ0.971 ± 0.0320.907 ± 0.0040.997 ± 0.129 0.983 ± 0.0010.991 ± 0.0010.980 ± 0.0000.987 ± 0.0010.999 ± 0.001
3WIZ0.973 ± 0.0190.906 ± 0.0070.998 ± 0.0110.977 ± 0.0000.999 ± 0.0010.975 ± 0.0000.988 ± 0.0050.999 ± 0.000
BRD4-15D3L0.962 ± 0.0280.972 ± 0.0070.968 ± 0.0130.936 ± 0.0010.952 ± 0.0010.917 ± 0.0010.947 ± 0.0130.999 ± 0.003
5KU30.972 ± 0.0490.966 ± 0.0050.957 ± 0.0080.924 ± 0.0010.927 ± 0.0010.900 ± 0.0010.928 ± 0.0100.999 ± 0.003
CREBBP5EIC0.635 ± 0.0420.500 ± 0.0000.887 ± 0.1780.794 ± 0.0010.938 ± 0.0030.766 ± 0.0000.766 ± 0.0110.938 ± 0.033
5MMG0.714 ± 0.0700.500 ± 0.0000.883 ± 0.1530.746 ± 0.0090.946 ± 0.0040.734 ± 0.0140.760 ± 0.0220.934 ± 0.028
HIV IN4CFD0.970 ± 0.0080.920 ± 0.0170.936 ± 0.0040.827 ± 0.0040.895 ± 0.0020.781 ± 0.0020.831 ± 0.0240.999 ± 0.005
4CHO0.931 ± 0.0550.928 ± 0.0090.929 ± 0.0190.832 ± 0.0140.896 ± 0.0030.770 ± 0.0010.833 ± 0.0520.999 ± 0.005
XIAP1TFT0.972 ± 0.0890.974 ± 0.0380.998 ± 0.1270.981 ± 0.0000.997 ± 0.0000.984 ± 0.0000.994 ± 0.0050.999 ± 0.000
5C3H0.985 ± 0.0060.984 ± 0.0070.998 ± 0.1160.981 ± 0.0030.998 ± 0.0000.987 ± 0.0000.997 ± 0.0000.999 ± 0.000
Mcl-15FC40.752 ± 0.068 0.578 ± 0.0060.930 ± 0.1430.811 ± 0.0020.835 ± 0.0020.817 ± 0.0010.842 ± 0.0100.956 ± 0.008
5MES0.780 ± 0.0890.583 ± 0.0030.921 ± 0.147 0.819 ± 0.0020.877 ± 0.0040.817 ± 0.0030.843 ± 0.0100.957 ± 0.014
Mdm24ODF0.943 ± 0.0030.938 ± 0.0060.977 ± 0.0770.932 ± 0.0000.966 ± 0.0000.942 ± 0.0000.952 ± 0.0050.999 ± 0.003
4ZFI0.952 ± 0.0090.937 ± 0.0050.977 ± 0.0590.926 ± 0.0020.962 ± 0.0000.942 ± 0.0000.956 ± 0.0060.999 ± 0.002
Menin5DB20.500 ± 0.0000.500 ± 0.0000.823 ± 0.1290.758 ± 0.0140.903 ± 0.0040.662 ± 0.0020.680 ± 0.0280.969 ± 0.038
6B410.538 ± 0.0120.500 ± 0.0000.828 ± 0.2420.760 ± 0.0250.897 ± 0.0080.617 ± 0.0800.679 ± 0.0280.971 ± 0.033
Table 4. ROC AUC values for ten PPI targets obtained by applying ML models to the test sets. The best AUC value corresponding to each target protein (across rows) is shown in bold.
Table 4. ROC AUC values for ten PPI targets obtained by applying ML models to the test sets. The best AUC value corresponding to each target protein (across rows) is shown in bold.
TargetPDB IDTreeBagged ForestRandom ForestBayesianSVMLogistic RegressionNeural NetNeural Net (Bagging)
BAZ2B4XUA0.5340.50.5370.530.5210.5320.530.54
5E730.5240.50.5460.5250.540.520.5380.553
Bcl-22O210.9660.9490.9870.9780.9790.9720.9810.985
4LVT0.9480.9480.9920.9820.9760.9740.9880.991
Bcl-xL3INQ0.9640.910.9890.9820.9830.9850.990.986
3WIZ0.9790.9160.9890.9750.9810.9860.9880.993
BRD4-15D3L0.9260.9260.9180.8970.9150.8930.9170.918
5KU30.9010.9010.9110.8860.9080.8870.9060.927
CREBBP5EIC0.6310.50.7690.7260.7240.7440.750.772
5MMG0.6890.50.6910.7120.7320.7210.7460.788
HIV IN4CFD0.8220.8270.8010.740.8050.7420.7840.832
4CHO0.7920.8070.7910.730.8160.7140.7530.824
XIAP1TFT0.970.9720.9960.9790.990.9810.9940.997
5C3H0.9730.9760.9940.9740.9940.9850.9950.995
Mcl-15FC40.7420.5860.8540.7830.7760.8130.8430.859
5MES0.7550.5730.8450.7860.7150.8090.8350.846
Mdm24ODF0.9470.9470.9580.9220.9570.9460.950.963
4ZFI0.9410.940.9490.9190.9540.9430.9510.961
Menin5DB20.50.50.6740.6220.6140.620.6460.646
6B410.5360.50.6690.6010.6080.6330.6560.637
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Singh, N.; Villoutreix, B.O. A Hybrid Docking and Machine Learning Approach to Enhance the Performance of Virtual Screening Carried out on Protein–Protein Interfaces. Int. J. Mol. Sci. 2022, 23, 14364. https://doi.org/10.3390/ijms232214364

AMA Style

Singh N, Villoutreix BO. A Hybrid Docking and Machine Learning Approach to Enhance the Performance of Virtual Screening Carried out on Protein–Protein Interfaces. International Journal of Molecular Sciences. 2022; 23(22):14364. https://doi.org/10.3390/ijms232214364

Chicago/Turabian Style

Singh, Natesh, and Bruno O. Villoutreix. 2022. "A Hybrid Docking and Machine Learning Approach to Enhance the Performance of Virtual Screening Carried out on Protein–Protein Interfaces" International Journal of Molecular Sciences 23, no. 22: 14364. https://doi.org/10.3390/ijms232214364

APA Style

Singh, N., & Villoutreix, B. O. (2022). A Hybrid Docking and Machine Learning Approach to Enhance the Performance of Virtual Screening Carried out on Protein–Protein Interfaces. International Journal of Molecular Sciences, 23(22), 14364. https://doi.org/10.3390/ijms232214364

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop