| Literature DB >> 35133138 |
Sami T Kurkinen1,2,3, Jukka V Lehtonen4,5, Olli T Pentikäinen1,2,3, Pekka A Postila1,2,3.
Abstract
Molecular docking is a key in silico method used routinely in modern drug discovery projects. Although docking provides high-quality ligand binding predictions, it regularly fails to separate the active compounds from the inactive ones. In negative image-based rescoring (R-NiB), the shape/electrostatic potential (ESP) of docking poses is compared to the negative image of the protein's ligand binding cavity. While R-NiB often improves the docking yield considerably, the cavity-based models do not reach their full potential without expert editing. Accordingly, a greedy search-driven methodology, brute force negative image-based optimization (BR-NiB), is presented for optimizing the models via iterative editing and benchmarking. Thorough and unbiased training, testing and stringent validation with a multitude of drug targets, and alternative docking software show that BR-NiB ensures excellent docking efficacy. BR-NiB can be considered as a new type of shape-focused pharmacophore modeling, where the optimized models contain only the most vital cavity information needed for effectively filtering docked actives from the inactive or decoy compounds. Finally, the BR-NiB code for performing the automated optimization is provided free-of-charge under MIT license via GitHub (https://github.com/jvlehtonen/brutenib) for boosting the success rates of docking-based virtual screening campaigns.Entities:
Mesh:
Substances:
Year: 2022 PMID: 35133138 PMCID: PMC8889583 DOI: 10.1021/acs.jcim.1c01145
Source DB: PubMed Journal: J Chem Inf Model ISSN: 1549-9596 Impact factor: 4.956
Figure 1Negative image-based rescoring. (A) Ligand binding cavity of RXRα (gray cartoon; PDB: 1MV9(23)) with a co-crystallized agonist (yellow stick model). (B) NIB model against the cavity cross-section (gray surface). The NIB model, depicting the key cavity shape/electrostatic potential features, is composed of negative (O; red), positive (N; blue), and neutral (C; black) cavity atoms (spheres). (C) Overlay of docking poses are shown for seven active ligands (magenta stick models), where (D) co-crystallized agonist is bound. (E) Best (green) and worst (red) shape/electrostatic matches for two docked actives (stick models) are shown with the NIB model (cyan surface). (F) Receiver operating characteristic (ROC) curves show that the R-NiB (red line) boosts the docking yield (blue line).
Figure 2Brute force negative image-based optimization. (A) During the first generation (Gen #1) of the optimization, each cavity atom is removed from the initial NIB model (Gen #0) one at the time to generate eight new seven-atom variants. If one of the variants improves enrichment (node boxed green) more than the other variants or the input in docking rescoring, it is used as the input for another round of editing/benchmarking. Here, the iteration goes through Gens #2–4 as the enrichment improves, and none of the last variants (Gen #5 = Gen #X) improves on the best model (Gen #4; node with a green background). (B) Atom composition evolution is shown for retinoid X receptor alpha for the input (Gen #0), mid-point (Gen #11), and fully optimized (Gen #23) NIB models. The rescoring with the final model (red line) is superior to docking (blue line), as seen in the ROC curves (x axis logarithmic). (C) Optimization protocol includes testing and potentially even validation prior to the virtual screening usage.
Docking and Brute Force Negative Image-Based Rescoring with the Test Setsa
| train/test | method | yield | COX2 | RXRα | MR | NEU | PDE5 | ER | PPARγ |
|---|---|---|---|---|---|---|---|---|---|
| 100:100 | docking | AUC | 0.66 ± 0.01 | 0.77 ± 0.02 | 0.55 ± 0.03 | 0.85 ± 0.02 | 0.78 ± 0.01 | 0.74 ± 0.01 | |
| EFd 1% | 5.7 | 11.5 | 3.2 | 4.1 | 11.3 | 21.7 | 24.2 | ||
| EFd 5% | 21.6 | 37.4 | 19.1 | 32.7 | 28.1 | 36.6 | |||
| BR20 | 0.22 | 0.35 | 0.17 | 0.29 | 0.28 | 0.36 | |||
| BR-NiB | AUC | ||||||||
| EFd 1% | |||||||||
| EFd 5% | 52.7 | ||||||||
| BR20 | 0.47 | ||||||||
| BR-NiB + shape only | AUC | 0.68 ± 0.02 | |||||||
| EFd 1% | |||||||||
| EFd 5% | |||||||||
| BR20 | |||||||||
| 70:30 | docking | AUC | 0.66 ± 0.02 | 0.77 ± 0.03 | 0.53 ± 0.04 | 0.85 ± 0.03 | 0.77 ± 0.02 | 0.77 ± 0.02 | |
| EFd 1% | 5.5 | 12.1 | 1.5 | 2.9 | 10.8 | 28.2 | 21.2 | ||
| EFd 5% | 19.5 | 37.4 | 15.4 | 32.4 | 27.0 | 34.8 | |||
| BR20 | 0.20 | 0.35 | 0.14 | 0.29 | 0.27 | 0.42 | 0.50 | ||
| BR-NiB | AUC | ||||||||
| EFd 1% | |||||||||
| EFd 5% | 49.3 | ||||||||
| BR20 | 0.46 | ||||||||
| BR-NiB + shape only | AUC | 0.69 ± 0.03 | |||||||
| EFd 1% | 28.2 | ||||||||
| EFd 5% | 54.1 | ||||||||
| BR20 | |||||||||
| 10:90 | docking | AUC | 0.67 ± 0.02 | 0.77 ± 0.03 | 0.56 ± 0.03 | 0.85 ± 0.03 | |||
| EFd 1% | 5.6 | 13.6 | 3.5 | 3.4 | 10.9 | 22.6 | 23.9 | ||
| EFd 5% | 22.2 | 39.8 | 17.6 | 31.5 | 28.4 | 36.8 | |||
| BR20 | 0.21 | 0.37 | 0.17 | 0.28 | 0.28 | 0.36 | |||
| BR-NiB | AUC | 0.75 ± 0.01 | 0.66 ± 0.02 | 0.81 ± 0.01 | |||||
| EFd 1% | 5.3 | 11.9 | |||||||
| EFd 5% | 20.1 | 42.4 | |||||||
| BR20 | 0.20 | 0.37 | |||||||
| BR-NiB + shape only | AUC | 0.68 ± 0.03 | 0.81 ± 0.01 | ||||||
| EFd 1% | |||||||||
| EFd 5% | 34.7 | 42.0 | |||||||
| BR20 | 0.36 | 0.42 |
The best values are underlined. The rescoring values are shown in bold and italics, if improved in comparison to the docking of each set (70, 30, 10, or 90%). Only those AUC values that are outside the error margin are highlighted. The Wilcoxon statistic[34] was used for the AUC error estimation.
Training/test set ratios (100:100, 70:30, and 10:90): the percentage of ligands used in the training (100, 70, and 10%) in relation to the percentage used in the testing (100, 30, and 90%).
Methods: flexible docking (PLANTS) and BR-NiB either with the equal shape/ESP (0.5/0.5) weight or the shape only (1.0/0.0).
Extra Test Set Results on Five DUD-E Targets and the Alternative PPARγ Structurea
| train/test | method | yield | AKT1 | DRD3 | ACES | COMT | FAK1 | PPARγ |
|---|---|---|---|---|---|---|---|---|
| docking | AUC | 0.70 ± 0.03 | 0.60 ± 0.03 | 0.39 ± 0.02 | 0.68 ± 0.08 | 0.80 ± 0.05 | ||
| EFd 1% | 5.7 | 0.7 | 0.0 | 7.7 | 20.0 | 18.4 | ||
| EFd 5% | 22.7 | 6.3 | 2.9 | 23.1 | 36.7 | 46.9 | ||
| BR20 | 0.25 | 0.06 | 0.03 | 0.22 | 0.38 | 0.42 | ||
| BR-NiB | AUC | 0.67 ± 0.08 | ||||||
| EFd 1% | ||||||||
| EFd 5% | ||||||||
| BR20 |
The best values are shown in bold and italics for the test sets (30%). Only those AUC values that are outside the error margin are highlighted. The Wilcoxon statistic[32] was used for the AUC error estimation. An alternative protein 3D structure (PDB: 2GTK) was used for PPARγ docking and BR-NiB optimization.
Training/test set ratio (70:30): the percentage of ligands used in the training (70%) in relation to the percentage used in the testing (30%).
Methods: flexible docking (PLANTS) and BR-NiB with the equal shape/ESP (0.5/0.5) weight.
The shape only (1.0/0.0 of shape/ESP) weight of scoring results is shown in parentheses for PPARγ.
Figure 3Fusing negative image-based models to boost BR-NiB for PDE5. If R-NiB (Figure ) was done using an NIB model focusing on sildenafil- (Model I; orange line; PDB: 1UDT(35)) or tadalafil-specific (Model II; cyan line; PDB: 1XOZ(36)) binding volume, the PLANTS scoring (black line) worked better in comparison. In fact, ROC curves indicate that the R-NiB treatment worsened the yield. The fusion of these models lowered the enrichment even further (green line; Gen #0); however, the BR-NiB (magenta line; Gen #72) of the hybrid model boosted the docking performance substantially. See Figure for interpretation.
Figure 4Effect of docking software on the BR-NiB with NEU. The NIB model of NEU binding cavity (NEU; PDB: 1B9V(38)) was optimized using the docking poses of PLANTS, GOLD, GLIDE SP, or DOCK with BR-NiB (Figure ; Videos S1 and S2). The optimized models are alike; however, the composition differences due to alternative docking sampling affect the ROC curves.
Test Set Results of Five DUDE-Z and Extrema Setsa
| set | software/method | yield | NEU | AA2AR | ACES | HSP90 | AR |
|---|---|---|---|---|---|---|---|
| docking 70:30 | AUC | 0.90 ± 0.03 | 0.75 ± 0.02 | 0.33 ± 0.02 | 0.52 ± 0.05 | 0.60 ± 0.03 | |
| EFd 1% | 29.0 | 27.4 | 1.4 | 0.0 | 1.2 | ||
| EFd 5% | 58.1 | 41.5 | 4.3 | 3.2 | 13.3 | ||
| BR20 | 0.53 | 0.41 | 0.07 | 0.05 | 0.13 | ||
| BR-NiB 70:30 | AUC | 0.95 ± 0.02 | 0.56 ± 0.05 | ||||
| EFd 1% | 23.8 | ||||||
| EFd 5% | |||||||
| BR20 | |||||||
| training set ligs/decs | 66/4270 | 352/20,402 | 300/6690 | 67/3990 | 188/9625 | ||
| test set ligs/decs | 31/1830 | 164/8742 | 140/2868 | 31/1710 | 83/4123 | ||
| docking70:30 | AUC | 0.80 ± 0.05 | 0.93 ± 0.01 | 0.55 ± 0.03 | 0.47 ± 0.05 | 0.56 ± 0.03 | |
| EFd 1% | 10.0 | 46.9 | 3.7 | 0.0 | 0.0 | ||
| EFd 5% | 13.3 | 66.2 | 10.3 | 3.7 | 0.0 | ||
| BR20 | 0.13 | 0.63 | 0.10 | 0.03 | 0.02 | ||
| BR-NiB 70:30 | AUC | ||||||
| EFd 1% | |||||||
| EFd 5% | |||||||
| BR20 | |||||||
| training set ligs/decs | 68/48,588 | 337/94,790 | 317/65,995 | 61/76,469 | 188/98,175 | ||
| test set ligs/decs | 30/20,823 | 145/40,769 | 136/28,281 | 27/32,775 | 81/42,072 | ||
The values are shown in bold and italics, if improved in comparison to the docking. Only those AUC values that are over the error margin are highlighted. The testing was performed with the 70:30 training/test set division using equal shape/ESP (0.5/0.5) weight. The Wilcoxon statistic[34] was used for the AUC error estimation. The PDB codes for the target structures used are the following: 1B9V (NEU), 3EML (AA2AR), 6LTK (HSP90), 2AM9 (AR), and 1 × 1066 (ACES).
The DUDE-Z is the optimized version of the DUD-E set.
The active compounds from the DUD-E set were used (training/test) in addition to the Extrema decoys.
Docking and Brute Force Negative Image-Based Rescoring with Validation Setsa
| docking | BR-NiB-guided rescoring | ||||||
|---|---|---|---|---|---|---|---|
| target | yield | IC50 < 1 μM | IC50 < 50 μM | IC50 1–50 μM | IC50 < 1 μM | IC50 < 50 μM | IC50 1–50 μM |
| MR | AUC | 0.37 ± 0.07 | 0.44 ± 0.06 | 0.63 ± 0.07 | |||
| EFd 0.1% | 0 | 0 | 0 | ||||
| EFd 0.5% | 0 | 5 | 0 | ||||
| EFd 1% | 0 | 10 | 5 | ||||
| EFd 5% | 5 | 10 | 5 | ||||
| NEU | AUC | ||||||
| EFd 0.1% | 0 | 5 | 5 | ||||
| EFd 0.5% | 15 | 10 | 5 | ||||
| EFd 1% | 20 | 15 | 15 | ||||
| EFd 5% | 65 | 70 | 70 | ||||
Docking with default scoring (PLANTS) and BR-NiB-guided rescoring (SHAEP) were tested with three validation sets: “high potency” (IC50 ≤ 1 μM), “low-to-mid potency” (IC50 = 1–50 μM), or “high-to-low potency” (IC50 = ≤ 50 μM) for MR and NEU. The active ligands (N = 20) were mixed into the SPECS library (N = 140,626), making the active compound concentration 0.014% for the validation sets. The best results for each potency level are shown in bold and italics. The Wilcoxon statistic[34] was used for the AUC error estimation.