| Literature DB >> 34723543 |
Dong Chen1,2, Jiaxin Zheng1, Guo-Wei Wei2,3,4, Feng Pan1.
Abstract
The construction of appropriate representations remains essential for molecular predictions due to intricate molecular complexity. Additionally, it is often expensive and ethically constrained to generate labeled data for supervised learning in molecular sciences, leading to challenging small and diverse data sets. In this work, we develop a self-supervised learning approach to pretrain models from over 700 million unlabeled molecules in multiple databases. The intrinsic chemical logic learned from this approach enables the extraction of predictive representations from task-specific molecular sequences in a fine-tuned process. To understand the importance of self-supervised learning from unlabeled molecules, we assemble three models with different combinations of databases. Moreover, we propose a protocol based on data traits to automatically select the optimal model for a specific task. To validate the proposed method, we consider 10 benchmarks and 38 virtual screening data sets. Extensive validation indicates that the proposed method shows superb performance.Entities:
Year: 2021 PMID: 34723543 PMCID: PMC9358546 DOI: 10.1021/acs.jpclett.1c03058
Source DB: PubMed Journal: J Phys Chem Lett ISSN: 1948-7185 Impact factor: 6.888