| Literature DB >> 32708785 |
Eugene Lin1,2,3, Chieh-Hsin Lin3,4,5, Hsien-Yuan Lane3,6,7,8.
Abstract
A growing body of evidence now suggests that artificial intelligence and machine learning techniques can serve as an indispensable foundation for the process of drug design and discovery. In light of latest advancements in computing technologies, deep learning algorithms are being created during the development of clinically useful drugs for treatment of a number of diseases. In this review, we focus on the latest developments for three particular arenas in drug design and discovery research using deep learning approaches, such as generative adversarial network (GAN) frameworks. Firstly, we review drug design and discovery studies that leverage various GAN techniques to assess one main application such as molecular de novo design in drug design and discovery. In addition, we describe various GAN models to fulfill the dimension reduction task of single-cell data in the preclinical stage of the drug development pipeline. Furthermore, we depict several studies in de novo peptide and protein design using GAN frameworks. Moreover, we outline the limitations in regard to the previous drug design and discovery studies using GAN models. Finally, we present a discussion of directions and challenges for future research.Entities:
Keywords: artificial intelligence; de novo peptide and protein design; deep learning; dimension reduction; drug design; generative adversarial networks; machine learning; molecular de novo design; single-cell RNA sequencing
Mesh:
Substances:
Year: 2020 PMID: 32708785 PMCID: PMC7397124 DOI: 10.3390/molecules25143250
Source DB: PubMed Journal: Molecules ISSN: 1420-3049 Impact factor: 4.411
Figure 1An example of the generative adversarial network (GAN) architecture. The GAN architecture comprises two main components including a generative network module and a discriminative network module. Step 1: The generative network module produces synthetic instances as real as possible. Gaussian random noises normally serve as the input for the generative network module. One particular example in drug design and discovery is a reconstructed drug-like compound as a fake instance. Step 2: The discriminative network module assesses the probability that an instance stems from the real dataset. One particular example in drug design and discovery is a drug-like compound dataset. Step 3: Both the generative and discriminative network modules play concurrently against each other to obtain their objectives.
Figure 2An example of the deep adversarial autoencoder structure. The deep adversarial autoencoder structure comprises two main components including an autoencoder module and an adversarial network module. The autoencoder module comprises an encoder unit and a decoder unit. The encoder unit also serves as the generative network module of the adversarial network architecture. Step 1: The encoder unit produces synthetic instances as real as possible. One particular example in drug design and discovery is a reconstructed latent vector as a fake instance. Step 2: The discriminative network module assesses the probability that an instance stems from the real dataset. One particular example in drug design and discovery is a real latent vector from the drug-like compound dataset. Step 3: Both the autoencoder and discriminative network modules play concurrently against each other to obtain their objectives.
Relevant studies on the GAN-based structures of molecular de novo design.
| Study | Structure | Architecture | Object Generated | Learning Technique | Databases | Results |
|---|---|---|---|---|---|---|
| Kadurin et al. [ | druGAN | AAE | latent vector | autoencoder | PubChem | druGAN generated novel molecular compounds which can be considered as potential anticancer agents. |
| Guimaraes et al. [ | ORGAN | GAN | SMILES | RL | ZINC, | ORGAN performed better than recurrent neural networks or GAN alone. |
| Sanchez-Lengeling et al. [ | ORGANIC | GAN | SMILES | RL | ZINC, | ORGANIC showed good performance in terms of the quantitative estimate of drug-likeness, but not the Lipinski’s Rule-of-Five. |
| Putin et al. [ | RANC | GAN | SMILES | RL | ZINC, ChemDiv | RANC was superior to ORGANIC in terms of several drug discovery metrics. |
| Putin et al. [ | ATNC | GAN | SMILES | RL | ChemDiv | ATNC performed better than ORGANIC in terms of various functions. |
| Polykovskiy et al. [ | ECAAE | AAE | latent vector | autoencoder | ZINC | ECAAE generated novel molecular compounds which can be considered as target drugs in rheumatoid arthritis, psoriasis, and vitiligo. |
| Cao and Kipf [ | MolGAN | GAN | graph | RL | QM9 | MolGAN outperformed ORGAN and variational autoencoder-based structures. |
| Guarino et al. [ | DiPol-GAN | GAN | graph | RL | QM9 | DiPol-GAN had 1.3 times higher drug-likeliness scores than MolGAN. |
| Prykhodko et al. [ | LatentGAN | GAN | SMILES | autoencoder | ChEMBL | LatentGAN created novel drug-like compounds and was compatible to recurrent neural networks. |
| Maziarka et al. [ | Mol-CycleGAN | GAN | latent vector | direct flow | ZINC, ChEMBL | Mol-CycleGAN outperformed the junction tree variational autoencoder and the graph convolutional policy network structures. |
| Méndez-Lucio et al. [ | Conditioned GAN | GAN | latent vector | direct flow | L1000 | Conditioned GAN produced molecular compounds with desired gene expression signatures. |
AAE = adversarial autoencoder; ATNC = Adversarial Threshold Neural Computer; druGAN = drug Generative Adversarial Network; ECAAE = Entangled Conditional Adversarial AutoEncoder; GAN = Generative Adversarial Network; LatentGAN = Latent Generative Adversarial Networks; MolGAN = Molecular Generative Adversarial Network; Mol-CycleGAN = Molecular Cycle Generative Adversarial Network; ORGAN = Objective-Reinforced Generative Adversarial Networks; ORGANIC = Objective-Reinforced Generative Adversarial Network for Inverse-design Chemistry; RANC = Reinforced Adversarial Neural Computer.
Figure 3An example of a workflow of the generative adversarial network (GAN) architecture for molecular de novo design. Step 1: The generative network module produces synthetic drug-like compounds (which are generated as latent vectors, SMILES, or graphs) as real as possible. Step 2: The discriminative network module assesses the probability that a drug-like compound stems from the real drug-like compound datasets (for example, ChEMBL). Step 3: Both the generative and discriminative network modules play concurrently against each other to obtain their objectives. Note that the solutions to provide a flow for gradients include reinforcement learning techniques, an autoencoder module, and a direct flow.
Figure 4An example of the DR-A (Dimensionality Reduction with Adversarial variational autoencoder) model for dimensionality reduction in scRNA-seq analysis. Step 1: The encoder unit produces synthetic latent vectors as real as possible. The encoder unit provides the mean and covariance of the Gaussian distribution to serve as the variational distribution, which is commonly generated by a variational autoencoder structure. Step 2: On the other hand, the decoder unit produces reconstructed scRNA-seq data as real as possible. Step 3: The DR-A model has two discriminative network modules. The first discriminative network module assesses the probability that the latent vector stems from the real latent vectors. Step 4: The second discriminative network module assesses the probability that the scRNA-seq data stems from the real scRNA-seq datasets. (e) Step 5: The autoencoder and two discriminative network modules play concurrently against each other to obtain their objectives.
Figure 5An example of the deep adversarial variational autoencoder structure for dimensionality reduction in scRNA-seq analysis. Step 1: The encoder unit produces synthetic latent vectors as real as possible. The encoder unit provides the mean and covariance of the Gaussian distribution to serve as the variational distribution, which is commonly generated by a variational autoencoder structure. Step 2: The discriminative network module assesses the probability that the latent vector stems from the real latent vectors. Step 3: Both the autoencoder and discriminative network modules play concurrently against each other to obtain their objectives.