Literature DB >> 33743817

Molecular optimization by capturing chemist's intuition using deep neural networks.

Jiazhen He1, Huifang You2,3, Emil Sandström2,4, Eva Nittinger5, Esben Jannik Bjerrum2, Christian Tyrchan5, Werngard Czechtizky5, Ola Engkvist2.   

Abstract

A main challenge in drug discovery is finding molecules with a desirable balance of multiple properties. Here, we focus on the task of molecular optimization, where the goal is to optimize a given starting molecule towards desirable properties. This task can be framed as a machine translation problem in natural language processing, where in our case, a molecule is translated into a molecule with optimized properties based on the SMILES representation. Typically, chemists would use their intuition to suggest chemical transformations for the starting molecule being optimized. A widely used strategy is the concept of matched molecular pairs where two molecules differ by a single transformation. We seek to capture the chemist's intuition from matched molecular pairs using machine translation models. Specifically, the sequence-to-sequence model with attention mechanism, and the Transformer model are employed to generate molecules with desirable properties. As a proof of concept, three ADMET properties are optimized simultaneously: logD, solubility, and clearance, which are important properties of a drug. Since desirable properties often vary from project to project, the user-specified desirable property changes are incorporated into the input as an additional condition together with the starting molecules being optimized. Thus, the models can be guided to generate molecules satisfying the desirable properties. Additionally, we compare the two machine translation models based on the SMILES representation, with a graph-to-graph translation model HierG2G, which has shown the state-of-the-art performance in molecular optimization. Our results show that the Transformer can generate more molecules with desirable properties by making small modifications to the given starting molecules, which can be intuitive to chemists. A further enrichment of diverse molecules can be achieved by using an ensemble of models.

Entities:  

Keywords:  ADMET; Matched molecular pairs; Molecular optimization; Recurrent neural networks; Seq2seq; Transformer

Year:  2021        PMID: 33743817      PMCID: PMC7980633          DOI: 10.1186/s13321-021-00497-0

Source DB:  PubMed          Journal:  J Cheminform        ISSN: 1758-2946            Impact factor:   5.514


  3 in total

1.  Siamese Recurrent Neural Network with a Self-Attention Mechanism for Bioactivity Prediction.

Authors:  Daniel Fernández-Llaneza; Silas Ulander; Dea Gogishvili; Eva Nittinger; Hongtao Zhao; Christian Tyrchan
Journal:  ACS Omega       Date:  2021-04-15

2.  Transformer-based molecular optimization beyond matched molecular pairs.

Authors:  Jiazhen He; Eva Nittinger; Christian Tyrchan; Werngard Czechtizky; Atanas Patronov; Esben Jannik Bjerrum; Ola Engkvist
Journal:  J Cheminform       Date:  2022-03-28       Impact factor: 5.514

3.  Augmented Hill-Climb increases reinforcement learning efficiency for language-based de novo molecule generation.

Authors:  Morgan Thomas; Noel M O'Boyle; Andreas Bender; Chris de Graaf
Journal:  J Cheminform       Date:  2022-10-03       Impact factor: 8.489

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.