Literature DB >> 32358545

The ANI-1ccx and ANI-1x data sets, coupled-cluster and density functional theory properties for molecules.

Justin S Smith1,2, Roman Zubatyuk2,3, Benjamin Nebgen2, Nicholas Lubbers4, Kipton Barros4, Adrian E Roitberg5, Olexandr Isayev6, Sergei Tretiak7.   

Abstract

Maximum diversification of data is a central theme in building generalized and accurate machine learning (ML) models. In chemistry, ML has been used to develop models for predicting molecular properties, for example quantum mech<span class="Gene">anics (QM) calculated potential energy surfaces and atomic charge models. The <class="Chemical">span class="Chemical">ANI-1x and ANI-1ccx ML-based general-purpose potentials for organic molecules were developed through active learning; an automated data diversification process. Here, we describe the ANI-1x and ANI-1ccx data sets. To demonstrate data diversity, we visualize it with a dimensionality reduction scheme, and contrast against existing data sets. The ANI-1x data set contains multiple QM properties from 5 M density functional theory calculations, while the ANI-1ccx data set contains 500 k data points obtained with an accurate CCSD(T)/CBS extrapolation. Approximately 14 million CPU core-hours were expended to generate this data. Multiple QM calculated properties for the chemical elements C, H, N, and O are provided: energies, atomic forces, multipole moments, atomic charges, etc. We provide this data to the community to aid research and development of ML models for chemistry.

Entities:  

Year:  2020        PMID: 32358545      PMCID: PMC7195467          DOI: 10.1038/s41597-020-0473-z

Source DB:  PubMed          Journal:  Sci Data        ISSN: 2052-4463            Impact factor:   6.444


Background & Summary

Machine learning (ML) and data driven methods have far reaching applications across much of science and engineering. The power of machine learning stems from its ability to generalize predictions to unseen samples. Formulation of accurate and general-use ML models, especially neural network models, requires data sets that maximally span the problem space of interest. As real world examples of this need, data sets for autonomous vehicles require data from varying geographic locations, diverse weather conditions, urban and rural areas, while data sets for teaching autonomous drones to fly require crash scenarios to perform well[1]. A grand challenge for the machine learning community is determining the best way to develop sufficiently diverse data sets. Robotics has spearheaded the efforts to build such data sets through the use of active learning[2]: building data sets by asking ML models to choose what data needs to be added to a training set to perform better next time. Although the conce<span class="Chemical">pt of active learning originates from robotics, it has recently grown into an extremely important tool for collecting quantum chemistry data sets for use in ML applications[3-11]. Building an o<span class="Chemical">ptimally diverse data set for training ML models is a problem-class="Chemical">specific task. In this work, we define diversity in data as an adequate coverage of the problem class="Chemical">space of interest that allows ML algorithms to generalize to unseen samples from many sources. In chemistry, diverse data sets can be difficult to come by. For example, ML has been applied for the prediction of chemical reactions based on experimental data, but models can be biased due to the lack of available data where reactions fail[12]. Such biases can be attributed to a lack of diversity in the training data. Herr et al. show standard molecular dynamics sampling is an inefficient method for sampling in the development of ML potentials because it leads to a lack of diversity in data[13]. An important task for any researcher applying ML is to determine what represents a sufficiently diverse data set to build general use models in their domain. In chemistry, ML methods have been developed for the prediction of ab initio quantum mech<span class="Gene">anics (QM) computed properties, e.g. molecular potential energies and forces[14-18], atomic charges[19-21], molecular dipoles[22], HOMO-LUMO gaps, and more[23,24]. ML has also been applied for the prediction of QM-based Hamiltonians[25,26], which provide the electronic structure for a given system. Machine learning approaches promise to revolutionize the computational study of atomic systems by providing a highly accurate and computationally efficient route for obtaining QM accurate properties. These methods are typically thousands to millions of times faster than reference electronic structure methods, which use the most advanced numerical algorithms. ML methods for QM property prediction utilize various techniques including kernel ridge regression (KRR)[27,28], Gaussian process regression (GPR)[29,30], neural networks[15,18,31-34] and others[35-37]. Several benchmark data sets for measuring the prediction accuracy of ML-based QM property predictors have been publicly released. The QM9[23] and QM7[24] data sets have become staple benchmarks in the field. The QM7 benchmark consists of 7 k molecules with their associated energies while QM9 consists of multiple QM properties from up to 130 k different molecules computed with density functional theory (DFT). A subset of QM9 was also computed with highly accurate G4class="Gene">MP2 calculations. To date, many ML methods have achieved mean absolute errors (MAE) for molecular potential energy prediction on the order of 0.2 kcal/mol on the QM9 benchmark[16,18,38], with 1 kcal/mol being considered the “gold standard” of chemical accuracy. These benchmarks have led to many successes in the field of QM property prediction, eclass="Chemical">specially considering early models had MAE on the order of 5 to 10 kcal/mol on the QM7 benchmark[27]. The recently released Alchemy data set aims to extend the molecular size of data in the QM9 data set from 9 non-<class="Chemical">span class="Chemical">hydrogen atoms to 14[39]. The QM7, QM9 and similar benchmarks aim to train predictive models of QM properties, such as potential energies, at energy minimized molecular geometries computed using DFT. Since these data sets only span chemical space but the structures are at their equilibrium ground state geometries, ML models trained to them are only shown to provide accurate predictions for molecular structures obtained with an expensive DFT geometry optimization. Recent work by Lu et al. aims to alleviates this problem using force field optimized geometries and transfer learning[40]. Other benchmarks utilize single molecule MD trajectories[15]. The goal of single molecule benchmarks is to measure the accuracy of an ML model for learning the local potential surface of a given molecule. While both types of aforementioned benchmarks, equilibrium molecules and single molecule potential surface, have been pioneers in the field by providing a basis of comparison for different models, neither has aimed to cover both the chemical (connectivity of different elements) and conformational (same connectivity but differing geometries) spaces of molecules, simultaneously. More recent research has focused on the construction of data sets with millions of highly diverse data points with the goal of creating a general-purpose ML-based potential[18,41,42]. The data for training such highly flexible ML models must contain the necessary information for predicting a complete potential energy surface for a class of molecules. Constructing these data sets represents a challenging case since the dimensionality of the problem is very high, and the conformational space of a molecule is not known a priori. The class="Chemical">ANI-1 data set[43], which consists of over 20 million conformations derived from 57 thousand distinct molecular configurations containing the C, H, N, and O chemical elements, is an example of such a data set. This data set was used to build the general-purpose <class="Chemical">span class="Chemical">ANI-1 potential, which was shown to accurately predict the non-equilibrium potential surface of molecules much larger than those included in the training data set. The ISO-17 benchmark[15], another example of a benchmark that spans both chemical and conformational space, contains 5000 molecular conformations extracted from finite temperature MD simulations of 129 isomers with the chemical formula C7O2H10. Finally, the ChemSpider data set[42] is comprised of potential energies and forces of 3 million conformations from 15 thousand different C, H, N, and O containing molecules from the ChemSpider database. In this data descriclass="Chemical">ptor, we provide the <class="Chemical">span class="Chemical">ANI-1x[5] and ANI-1ccx[44] data sets to the broader scientific community. The ANI-1x data set contains DFT calculations for approximately five million diverse molecular conformations. These conformations were obtained through an active learning algorithm, whereby the ANI ML potential is trained in an iterative manner to determine what new data should be included in future versions of the training data set. The fact that the ANI-1x and ANI-1ccx data sets are generated through non-equilibrium sampling rather than equilibrium sampling means that these data set must contain a more diverse set of conformations than existing QMx style data sets. The ANI-1x data set also contains vastly more diverse molecular conformations than those supplied in the original ANI-1 random sampled data set since active learning methods were employed in its construction. After many iterations of active learning, the resulting data set, dubbed the ANI-1x data set, was employed to train the ANI-1x potential. The ANI-1ccx data set is an intelligently selected 10% sub-sample of the ANI-1x data set, but recomputed with an accurate coupled cluster (approximately CCSD(T)/CBS) level of theory. Portions of the ANI-1x data set were recomputed with a larger basis set using the wb97x DFT functional for later work[31]. The data sets are provided in one easy to access HDF5 file along with scripts and examples needed to extract the data. The combination of all aforementioned DFT and coupled cluster variants of the ANI-1x data sets, with a selection of properties, are supplied to the scientific community in this data descriptor. Future extensions of these data sets will add new chemical elements and more molecular diversity will also be released to the community[45].

Methods

These methods are expanded versions of descri<span class="Chemical">ptions in our related work on developing general-purpose machine learning potentials for org<class="Chemical">span class="Gene">anic molecules through active learning[5] and transfer learning[44].

Active learning data selection

The initial class="Chemical">ANI-1x data set was generated as a part of an active learning procedure to develop the <class="Chemical">span class="Chemical">ANI-1x potential[5]. Active learning is where an ML model is used to determine what new data should be included in later generations to improve predictive ability. Figure 1a depicts the active learning algorithm. First, an ensemble of ANI models is trained to an initial bootstrap data set. Databases of molecules such as GDB-11[46,47], ChEMBL[48], and generated (with the RDKit cheminformatics python package[49]) amino acids and 2-amino acid peptides are randomly sampled for new molecule configurations, and one of four types of active learning sampling techniques is carried out on each of the selected molecules. These sampling techniques include molecular dynamics sampling, normal mode sampling, dimer sampling, and torsion sampling. These methods are described further in then Sampling methods section. Our active learning procedure involves a search through chemical and conformational space, employing a measure of estimated uncertainty to choose what new data should be generated, then including the new data in the next training cycle. The uncertainty estimate provides a priori information about the ensembles predictive performance. The uncertainty estimate employed in the ANI-1x active learning is based on an ensemble disagreement measure, henceforth referred to as ρ. The value ρ is proportional to the standard deviation of the prediction of an ensemble of ML models, normalized by the square root of the number of atoms in the system. It is described in detail in a previous publication[5]. When the uncertainty metric hints that a given molecular structure is poorly described (i.e., a large ρ value), new DFT data is generated and added to the training data set. The ensemble of models is then retrained with the new data added to the original data set. The new data is added in batches to accelerate the active learning process. The entire process is carried out iteratively to produce a successively more diverse data set, and hence a more robust ML model. In this work we use molecular dynamics (MD) simulations and geometry optimizations with our ML models. These simulations are performed with the Atomic Simulation Environment (ASE) python based library[50].
Fig. 1

Active learning schemes for building ANI data sets. (a) The active learning algorithm employed during the construction of the ANI-1x data set. (b) The ANI-1ccx selection and data generation scheme.

Active learning schemes for building <span class="Gene">ANI data sets. (a) The active learning algorithm employed during the construction of the <class="Chemical">span class="Chemical">ANI-1x data set. (b) The ANI-1ccx selection and data generation scheme.

Sampling methods

Molecular dynamics sampling

The molecular dynamics (MD) sampling method utilizes the currently trained ML model ensemble to drive an MD simulation of a randomly selected molecule. The MD simulation is carried out at a random temperature between 50 K and 800 K using the Langevin thermostat, a time step of 0.5 fs, and a friction coefficient of 0.02 fs−1. Every 5 MD time steps ρ is computed for the molecular structure x at that time step. If the measure ρ is larger than a predetermined value, x is selected and added to a set of high ρ molecules. The MD simulation is terminated once x is selected, because we assume the current potential no longer describes the atomic physics of the system at that point. DFT data is generated for all molecules in and added to the training data set.

Normal mode sampling

This sampling technique[41] begins with the QM generation of normal modes for tens of thousands of molecules. The molecules are selected from the GDB-11 and ChEMBL[48] databases by employing the ensemble disagreement value ρ. When a molecule is selected the geometry is o<span class="Chemical">ptimized to the local energy minima, then normal modes and harmonic force constants are calculated with QM and stored. During normal mode sampling the normal modes and harmonic force constants are used to generate a set of N conformations {x}, where i indexes the conformation. The conformations are generated by diclass="Chemical">splacing the atoms a random distance along each normal mode. All N conformations are then tested with the current <class="Chemical">span class="Gene">ANI ensemble to obtain ρ. All ρ larger than a predetermined threshold are selected then added to a set of high ρ molecular conformations. DFT data is generated for all molecules in and added to the training data set.

Dimer sampling

Dimer sampling starts with the random selection of a rectangular periodic cell size within the range of 20 to 30 angstroms. Molecules from a subset of the GDB-11 database are then selected randomly, with a higher probability of choosing molecules with less non-class="Chemical">hydrogen atoms. The selected molecules are then embedded within the ado<class="Chemical">span class="Chemical">pted periodic cell with random positions and rotation, ensuring that no two atoms in different molecules are within a distance of 1.5 Å. The atom density of the box is also randomly determined within a range. The current generation ANI potential is used to run an MD simulation on the constructed box of molecules. MD is carried out at a random temperature between 50 K and 600 K using the Langevin thermostat, a time step of 0.5 fs, and a friction coefficient of 0.02 . After 100 time steps, the box is decomposed into a complete set of dimer structures {x}, where i indexes the dimers. Only dimer structures with at least two atoms, one from each monomer, within a distance cutoff of 6 Å are selected. The ML model ensemble is then used to determine for each i dimer, and if the dimer has a value greater than a predetermined cutoff, the dimer is added to a set of high dimers. DFT data is generated for all molecules in and added to the training data set.

Torsion sampling

The torsion sampling active learning cycles are carried out after the original class="Chemical">ANI-1x data set is produced as a part of building the <class="Chemical">span class="Chemical">ANI-1ccx potential[44]. A SMILES[51] string is selected from a set of drug molecules containing relevant torsions[52]. RDKit is used to embed the molecules in 3D space and select a rotate-able bond, i.e. a torsion. The current generation ML potential is used to optimize the starting geometry, and carryout a relaxed scan, incremented by 10 degrees over the entire torsion profile. At each of the 36 scan steps, is measured. If is larger than a predetermined threshold the scan is halted and the structure x is stored. The scan is halted at the point where is high because we assume that the potential should not be trusted to continue performing the scan. Normal modes and harmonic force constants for x are generated using the current generation ANI potential ensemble. x is then randomly perturbed along those normal modes to generate four additional slightly perturbed conformations. This is done to sample the space around the valley of the torsion profile, rather than directly sampling the path through the valley. The latter could lead to a potential that poorly describes the overall shape of the reference QM potential. The four perturbed structures are added a set of high structures. DFT data are generated for all molecules in and added to the training data set. A total of 20 iterations of this torsional sampling protocol are carried out on all molecules in the Sellers et al. set[52].

High-throughput coupled cluster calculation scheme

ML potentials can archive errors lower than 1 kcal/mol compared to the underlying QM reference method[16,38,41]. As ML methods become more accurate at fitting to reference data, it becomes advantageous to develop data sets from more accurate levels of theory. An accurate and affordable QM method is the coupled cluster method with the inclusion of single, double and perturbative triple excitations, i.e., CCSD(T). This method is usually referred to as the “gold standard” of computational chemistry methods. Recently, an approximate DLPNO-CCSD(T)[53] method was developed and implemented in the ORCA software package[54]. This method has a linear scaling with system size and is 10–100 times less expensive for medium-sized molecules (10–30 atoms). The error of the DLPNO approximation is controlled in the ORCA package with NormalPNO and TightPNO oclass="Chemical">ption presets. The TightPNO setting usually introduces a very small error, compared to the error of the pristine CCSD(T) method itself[55]. Besides the electronic level of theory, an extrapolation towards the complete basis set (CBS) limit is also a requirement to obtain high accuracy reference QM data. A very accurate and computationally efficient composite extrapolation scheme developed by Hobza and Sponer[56] uses an <class="Chemical">span class="Gene">MP2/(aug-)cc-pV[TQ]Z extrapolation and CCSD(T) correlation energy correction calculated with the (aug-)cc-pVTZ basis set. We used a similar idea to construct a composite extrapolation scheme, which is 50 times faster than full CCSD(T)/CBS for an aspirin-size molecule, without considerable sacrifice of accuracy[44]. The components of our extrapolation scheme, dubbed CCSD(T)*/CBS, are the following: the effect of increasing the DLPNO approximation accuracy setting is estimated as the difference between TighPNO and class="Disease">NormalPNO CCSD(T) calculations with the cc-pVDZ basis set. The DLPNO-CCSD(T)/cc-pVTZ energy is calculated with a NormalPNO setting. The difference between cc-pVTZ and CBS correlation energy is estimated with an extrapolation of <class="Chemical">span class="Gene">MP2 energy using the cc-pVTZ and cc-pVQZ basis sets. The HF/CBS energy is estimated with an extrapolation of cc-pVTZ and cc-pVQZ energies. The final equation for our CCSD(T)*/CBS composite extrapolation scheme is:where E(N, Method) refers to the energy of ‘Method’ calculated with basis set of cardinal number N (2, 3 and 4 for cc-pVDZ, cc-pVTZ, and cc-pVQZ), NPNO and TPNO are Normal and Tight DLPNO respectively, and Extrap(N1/N2, Method) refers to the complete basis set extrapolation performed with formulas from Halkier[57] and Helgaker[58]:Parameters α = 5.46 and β = 3.05 optimized for cc-pV[TQ]Z extrapolation are used from a previous report[59].

CCSD(T)*/CBS data selection

Since our CCSD(T)*/CBS extrapolation scheme is significantly more computationally expensive than the original <span class="Chemical">ANI-1x DFT reference calculations, only a subset of the original data could be computed using the extrapolation scheme. Therefore, we used a subset sampling technique similar to the original active learning method to select what data to compute at the CCSD(T)*/CBS level of theory. Figure 1b depicts the subset selection technique; first, a uniformly random subset of the <class="Chemical">span class="Chemical">ANI-1x data set is selected, then CCSD(T)*/CBS reference data is generated. An ensemble of models is trained to this new CCSD(T)*/CBS data set. The ensemble disagreement value is generated for all remaining ANI-1x data points. A subset of the data with greater than a predetermined value is selected, and CCSD(T)*/CBS reference data is generated. This data is added to the CCSD(T)*/CBS training data set. The cycle was carried out four times to generate the final ANI-1ccx data set. Supplemental information Tables 1 and 2 provides counts for all data in the ANI-1ccx data set.

Data Records

All data records [10.6084/m9.figshare.c.4712477] are provided in a single HDF5[60] file which is hosted by figshare[61]. The provided HDF5 file can be iterated through using a python scriclass="Chemical">pt named “example_loader.py”, which we provide in a publicly accessible GitHub repository [https://github.com/aiqm/<class="Chemical">span class="Gene">ANI1x_datasets]. Using our extraction scripts and examples, all data can easily be accessed. The Usage Notes section provides more details about how to access the data from the provided HDF5 file. The property name along with the dictionary keys, units, data type, and NumPy array shape, are provided in Table 1. Tables S1 and S2 in the supplemental information contains counts for all data in this article. Three primary electronic structure methods were used to generate the data in this article: wB97x/6-31G*, wB97x/def2-TZVPP, and CCSD(T)*/CBS. Many secondary methods for the CCSD(T)*/CBS extrapolation are also included. All wB97x/6-31G* calculations were carried out in the Gaussian 09[62] electronic structure package, while all wB97x/def2-TZVPP and CCSD(T)*/CBS calculations were performed in the ORCA[54] software package. Table S3 contains a map between the key value and the level of theory at which the property was computed. In all, we include multiple types of energies, forces, electronic multipole moments, and charges obtained from the Hirshfeld and CM5 charge partitioning schemes. Minimal basis iterative stockholder partitioning (MBIS) scheme[63] implemented in the HORTON software library[64] was used to calculate atomic charges, volumes, and the magnitude of higher order atomic multipoles up to octupoles based on the wB97x/def2-TZVPP electron density.
Table 1

Data layout in the provided HDF5 file. Nc is the number of conformations and Na is the number of atoms.

PropertyKeyUnitsTypeShape
Atomic Positions‘coordinates’Åfloat32(Nc, Na, 3)
Atomic Numbers‘atomic_numbers’uint8(Na)
Total Energy‘wb97x_dz.energy’Hafloat64(Nc)
‘wb97x_tz.energy’
‘ccsd(t)_cbs.energy’
HF Energy‘hf_dz.energy’Hafloat64(Nc)
‘hf_tz.energy’
‘hf_qz.energy’
NPNO-CCSD(T)‘npno_ccsd(t)_dz.corr_energy’Hafloat64(Nc)
Correlation‘npno_ccsd(t)_tz.corr_energy’
Energy‘npno_ccsd(t)_qz.corr_energy’
MP2‘mp2_dz.corr_energy’Hafloat64(Nc)
Correlation‘mp2_tz.corr_energy’
Energy‘mp2_qz.corr_energy’
Atomic Forces‘wb97x_dz.forces’Ha/Åfloat32(Nc, Na, 3)
‘wb97x_tz.forces’
Molecular‘wb97x_dz.dipole’e Åfloat32(Nc, 3)
Electric‘wb97x_tz.dipole’
Moments‘wb97x_tz.quadrupole’e AA2(Nc, 6)
Atomic‘wb97x_dz.cm5_charges’efloat32(Nc, Na)
Charges‘wb97x_dz.hirshfeld_charges’
‘wb97x_tz.mbis_charges’
Atomic‘wb97x_tz.mbis_dipoles’a.u.float32(Nc, Na)
Electric‘wb97x_tz.mbis_quadrupoles’
Moments‘wb97x_tz.mbis_octupoles’
Atomic Volumes‘wb97x_tz.mbis_volumes’a.u.float32(Nc, Na)
Data layout in the provided HDF5 file. Nc is the number of conformations and Na is the number of atoms.

Technical Validation

Diversity comparison

Here, we compare the diversity of atomic environments sampled in class="Chemical">ANI-1x and <class="Chemical">span class="Chemical">ANI-1ccx to two prior data sets: QM9 and ANI-1. The QM9 data set[23] is a collection of 134 k molecules with up to 9 non-hydrogen atoms (C, N, O, and F) along with molecular properties (electronic and vibrational energies, Highest Occupied Molecular Orbitals (HOMO) and Lowest Unoccupied Molecular Orbital (LUMO) energies, dipole moments, etc.) calculated at the DFT optimized conformer geometry with the B3LYP/6-31G(2df,p) method. These molecules correspond to a subset of the GDB-17[65] chemical universe of 166 billion organic molecules. The QM9 data set represented a molecular property set of unprecedented size and consistency at the time of its inception. Subsequently, this data set has become very popular for training ML property predictors for organic molecules[66-75]. However, the principal disadvantage of the QM9 data set is that it is composed of only optimized molecular structures. ML models trained on the QM9 data set should not be used for non-equilibrium simulations, such as MD. Further, applications of an ML model trained to QM9 requires a DFT geometry optimization to achieve accurate predictions, since all geometries in the training set were DFT optimized geometries. The class="Chemical">ANI-1 data set[43] is the predecessor to the data sets presented in this work. The <class="Chemical">span class="Chemical">ANI-1 data set contains approximately 20 million randomly selected structures and and their corresponding DFT computed energies. Similar to the QM7 and QM9 data sets, the ANI-1 data set covers chemical degrees of freedom. However, most importantly, it also aims to describe conformational degrees of freedom by providing non-equilibrium molecular geometries for many different molecules. The nearly 20 million non-equilibrium geometries were generated by randomly deforming molecules along DFT computed vibrational normal modes. This process was carried out for approximately 50 k different molecules with up to 8 heavy (C, N, O) atoms. Training an ANI model to this data set resulted in the ANI-1 potential, a model capable of extending its predictive accuracy to molecules much larger than those in the training data set. However, the ANI-1 model was random sampled from an enumeration of small organic molecules from GDB-11[46,47], which left the data set clustered, yet sparse, in conformational space coverage. The ANI-1x and ANI-1ccx data sets were generated to provide a more diverse sampling of conformational space for small organic molecules. We use features of the environment of each atom to compare the chemical and conformational space coverage for the QM9, class="Chemical">ANI-1, <class="Chemical">span class="Chemical">ANI-1x and ANI-1ccx data sets. Instead of comparing the atomic environment vectors (AEVs) of the ANI[41] model, we use intermediate neural network activations of a trained model to compare the similarity of two chemical environments; specifically, we use the activation values that form the first layer of the ANI-1x[5] model. We use activation values rather than the descriptors (AEVs) since AEVs tend to be sparse, making it difficult to use simple L2 distance between vectors as a similarity metric. The first layer extracts the most useful information from the AEV and reduces the dimensionality of the atomic descriptor from 384 to 160 (H), 144 (C) and 128 (N and O). For visual comparison, we applied the parametric t-Distributed Stochastic Neighbor Embedding (pt-SNE) technique[76] which learns a neural network-based mapping from high-dimensional vectors to a 2-dimensional representation, while preserving both global and local structure of the data in high-dimensional space as much as possible. We used the complete ANI-1x data set to learn a pt-SNE mapping from the ANI-1x first layer activations to a 2D space using a deep neural network. Figure 2 provides a direct comparison of 2D embedding of atom environment features for the entire QM9 data set and random subsets of the ANI-1, ANI-1x, and ANI-1ccx data sets with the same number of atoms for each element. The 2D embeddings were learned from the ANI-1x data set. The colors correspond to the type and number of bonded neighbors as determined by the Openbabel software library[77], e.g. the ethanol molecule has two types of C atoms with environment HHHC and HHCO. We also provide an interactive version on our supporting GitHub repository [https://github.com/aiqm/ANI1x_datasets], with all labels corresponding to the colors used.
Fig. 2

2D parametric t-SNE embeddings. These embeddings are for the 1st layer of activations of the ANI-1x model for the complete QM9 data set and random subsets of the ANI-1, ANI-1x and ANI-1ccx data sets. The same number of atoms are compared for each element. The different colors correspond to the number and type of bonded neighbors.

2D parametric t-SNE embeddings. These embeddings are for the 1st layer of activations of the <span class="Chemical">ANI-1x model for the complete QM9 data set and random subsets of the <class="Chemical">span class="Chemical">ANI-1, ANI-1x and ANI-1ccx data sets. The same number of atoms are compared for each element. The different colors correspond to the number and type of bonded neighbors. Visual comparclass="Chemical">ison of the <class="Chemical">span class="Chemical">pt-SNE embeddings clearly show atoms cluster according to the type and the local chemical environment. Carbon, for example, has distinct clusters corresponding to sp3, sp2 and sp hybridization for both QM9 and ANI data sets. All four data sets have a similar number of clusters, reflecting the chemical diversity of the data sets. The ANI-1x and ANI-1ccx data sets have noticeably more diffuse clusters compared to the QM9 data set. This fact reflects greater coverage of conformational space. Comparing the random conformational sampling of the ANI-1 data set with the active learning generated ANI-1x data sets shows a visually similar trend. However, upon close inspection it is clear that the ANI-1x data has a more diverse coverage of chemical space, especially for N and O. The coverage of the ANI-1x and ANI-1ccx data sets is very similar, which is expected since the ANI-1ccx data set is a sub-sampling of the ANI-1x data set.

Energy and composition coverage

Figure 3a provides a histogram showing the distribution of energies in the data sets. The energies have had a constant shift removed, which is the sum of a linear fitting to the atomic elements. The figure also shows the energy range of applicability for models trained to this data set. The <span class="Chemical">ANI-1ccx data set has a similar distribution as the <class="Chemical">span class="Chemical">ANI-1x data set since ANI-1ccx is an active learning-based sub-sample of ANI-1x. Figure 3b is a histogram of the number of atoms (C, H, N, and O) per molecule in each data set. After making note of the logarithmic scale on the y axis, it is clear that the vast majority of data in these data sets is derived from molecules with less than 20 atoms in total. The ANI-1ccx data set tends to have smaller molecules than the ANI-1x data set because the coupled cluster calculations on larger molecules sometimes fail to complete due to computer memory limitations.
Fig. 3

ANI data set energy and size distribution. (a) A histogram of the potential energies in the ANI-1x and ANI-1ccx data sets with a linear fit per atomic element E removed. The bin width is 1 millihartree. (b) A histogram of the total number of atoms (including C, H, N, and O atoms) per molecule in the ANI-1x and ANI-1ccx data sets. The bin width is one.

class="Gene">ANI data set energy and size distribution. (a) A histogram of the potential energies in the <class="Chemical">span class="Chemical">ANI-1x and ANI-1ccx data sets with a linear fit per atomic element E removed. The bin width is 1 millihartree. (b) A histogram of the total number of atoms (including C, H, N, and O atoms) per molecule in the ANI-1x and ANI-1ccx data sets. The bin width is one.

Usage Notes

We provide python based tools for extracting these data sets, along with code examples, on a public supporting GitHub repository [https://github.com/aiqm/class="Gene">ANI1x_datasets]. Box 1 provides an example of a python scri<class="Chemical">span class="Chemical">pt for loading the ANI-1x wb97x/6-31G* data set from the provided ANI-1x HDF5 file. After importing the data loader in line 0, a function called ‘iter_data_buckets’ provides an iterator that can be used in a for loop (line 5) to access dictionaries containing all data for chemical isomers. Two arguments are passed to this function. The first argument is the path to the ani-1x HDF5 file (defined in line 2). The second is a list of keys describing what data will be loaded (defined in line 4). The script will only load conformers that share the requested data for property keys given in the ‘data_keys’ list. For example, if the ‘data_keys’ list contains two keys ‘wb97x_dz.energy’ and ‘ccsd(t)_cbs.energy’, then only conformers that share both energies will be loaded, approximately 500k structures. However, if one removes ‘ccsd(t)_cbs.energy’ from the list, then approximately 5 million structures will be loaded. Finally, the data (geometry coordinates, atomic numbers, and requested properties) stored in NumPy arrays can be accessed by requesting the specific key, as shown in lines 6–9. Note that the keys match those given in Table 1. 0) import dataloader as dl 1) 2) path_to_h5file = ‘PATH/TO/H5/FILE/<span class="Gene">ani1x.h5’ 3) 4) data_keys = [‘wb97x_dz.energy’,‘wb97x_dz.forces’] 5) for data in dl.iter_data_buckets(path_to_h5file,keys=data_keys): 6) X = data[‘coordinates’ 7) Z = data[‘atomic_numbers’] 8) E = data[‘wb97x_dz.energy’] 9) F = data[‘wb97x_dz.forces’] supplemental information
Measurement(s)Quantum Mechanics • energy • force • multipole moment • atomic charge
Technology Type(s)computational modeling technique
Factor Type(s)atom
Sample Characteristic - Environmentorganic molecule
  29 in total

1.  A neural network potential-energy surface for the water dimer based on environment-dependent atomic energies and charges.

Authors:  Tobias Morawietz; Vikas Sharma; Jörg Behler
Journal:  J Chem Phys       Date:  2012-02-14       Impact factor: 3.488

2.  A Density Functional Tight Binding Layer for Deep Learning of Chemical Hamiltonians.

Authors:  Haichen Li; Christopher Collins; Matteus Tanha; Geoffrey J Gordon; David J Yaron
Journal:  J Chem Theory Comput       Date:  2018-11-05       Impact factor: 6.006

3.  Less is more: Sampling chemical space with active learning.

Authors:  Justin S Smith; Ben Nebgen; Nicholas Lubbers; Olexandr Isayev; Adrian E Roitberg
Journal:  J Chem Phys       Date:  2018-06-28       Impact factor: 3.488

4.  Metadynamics for training neural network model chemistries: A competitive assessment.

Authors:  John E Herr; Kun Yao; Ryker McIntyre; David W Toth; John Parkhill
Journal:  J Chem Phys       Date:  2018-06-28       Impact factor: 3.488

5.  Discovering a Transferable Charge Assignment Model Using Machine Learning.

Authors:  Andrew E Sifain; Nicholas Lubbers; Benjamin T Nebgen; Justin S Smith; Andrey Y Lokhov; Olexandr Isayev; Adrian E Roitberg; Kipton Barros; Sergei Tretiak
Journal:  J Phys Chem Lett       Date:  2018-07-27       Impact factor: 6.475

6.  Transferability in Machine Learning for Electronic Structure via the Molecular Orbital Basis.

Authors:  Matthew Welborn; Lixue Cheng; Thomas F Miller
Journal:  J Chem Theory Comput       Date:  2018-08-08       Impact factor: 6.006

7.  Machine Learning of Partial Charges Derived from High-Quality Quantum-Mechanical Calculations.

Authors:  Patrick Bleiziffer; Kay Schaller; Sereina Riniker
Journal:  J Chem Inf Model       Date:  2018-03-07       Impact factor: 4.956

8.  Data-Driven Learning of Total and Local Energies in Elemental Boron.

Authors:  Volker L Deringer; Chris J Pickard; Gábor Csányi
Journal:  Phys Rev Lett       Date:  2018-04-13       Impact factor: 9.161

9.  Machine-learning-assisted materials discovery using failed experiments.

Authors:  Paul Raccuglia; Katherine C Elbert; Philip D F Adler; Casey Falk; Malia B Wenny; Aurelio Mollo; Matthias Zeller; Sorelle A Friedler; Joshua Schrier; Alexander J Norquist
Journal:  Nature       Date:  2016-05-05       Impact factor: 49.962

10.  Quantum-chemical insights from deep tensor neural networks.

Authors:  Kristof T Schütt; Farhad Arbabzadah; Stefan Chmiela; Klaus R Müller; Alexandre Tkatchenko
Journal:  Nat Commun       Date:  2017-01-09       Impact factor: 14.919

View more
  15 in total

1.  Δ-Quantum machine-learning for medicinal chemistry.

Authors:  Kenneth Atz; Clemens Isert; Markus N A Böcker; José Jiménez-Luna; Gisbert Schneider
Journal:  Phys Chem Chem Phys       Date:  2022-05-11       Impact factor: 3.945

2.  Deep learning of dynamically responsive chemical Hamiltonians with semiempirical quantum mechanics.

Authors:  Guoqing Zhou; Nicholas Lubbers; Kipton Barros; Sergei Tretiak; Benjamin Nebgen
Journal:  Proc Natl Acad Sci U S A       Date:  2022-07-01       Impact factor: 12.779

Review 3.  Ab Initio Machine Learning in Chemical Compound Space.

Authors:  Bing Huang; O Anatole von Lilienfeld
Journal:  Chem Rev       Date:  2021-08-13       Impact factor: 60.622

4.  Linear Atomic Cluster Expansion Force Fields for Organic Molecules: Beyond RMSE.

Authors:  Dávid Péter Kovács; Cas van der Oord; Jiri Kucera; Alice E A Allen; Daniel J Cole; Christoph Ortner; Gábor Csányi
Journal:  J Chem Theory Comput       Date:  2021-11-04       Impact factor: 6.006

5.  Artificial intelligence-enhanced quantum chemical method with broad applicability.

Authors:  Peikun Zheng; Roman Zubatyuk; Wei Wu; Olexandr Isayev; Pavlo O Dral
Journal:  Nat Commun       Date:  2021-12-02       Impact factor: 14.919

Review 6.  Artificial intelligence and machine learning approaches for drug design: challenges and opportunities for the pharmaceutical industries.

Authors:  Chandrabose Selvaraj; Ishwar Chandra; Sanjeev Kumar Singh
Journal:  Mol Divers       Date:  2021-10-23       Impact factor: 2.943

7.  Combining Machine Learning and Computational Chemistry for Predictive Insights Into Chemical Systems.

Authors:  John A Keith; Valentin Vassilev-Galindo; Bingqing Cheng; Stefan Chmiela; Michael Gastegger; Klaus-Robert Müller; Alexandre Tkatchenko
Journal:  Chem Rev       Date:  2021-07-07       Impact factor: 60.622

8.  Dataset Construction to Explore Chemical Space with 3D Geometry and Deep Learning.

Authors:  Jianing Lu; Song Xia; Jieyu Lu; Yingkai Zhang
Journal:  J Chem Inf Model       Date:  2021-03-08       Impact factor: 4.956

9.  Complex reaction processes in combustion unraveled by neural network-based molecular dynamics simulation.

Authors:  Jinzhe Zeng; Liqun Cao; Mingyuan Xu; Tong Zhu; John Z H Zhang
Journal:  Nat Commun       Date:  2020-11-11       Impact factor: 14.919

10.  Multitask machine learning models for predicting lipophilicity (logP) in the SAMPL7 challenge.

Authors:  Eelke B Lenselink; Pieter F W Stouten
Journal:  J Comput Aided Mol Des       Date:  2021-07-17       Impact factor: 3.686

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.