Literature DB >> 34278230

Atom-specific persistent homology and its application to protein flexibility analysis.

David Bramer1, Guo-Wei Wei2.   

Abstract

Recently, persistent homology has had tremendous success in biomolecular data analysis. It works by examining the topological relationship or connectivity of a group of atoms in a molecule at a variety of scales, then rendering a family of topological representations of the molecule. However, persistent homology is rarely employed for the analysis of atomic properties, such as biomolecular flexibility analysis or B-factor prediction. This work introduces atom-specific persistent homology to provide a local atomic level representation of a molecule via a global topological tool. This is achieved through the construction of a pair of conjugated sets of atoms and corresponding conjugated simplicial complexes, as well as conjugated topological spaces. The difference between the topological invariants of the pair of conjugated sets is measured by Bottleneck and Wasserstein metrics and leads to an atom-specific topological representation of individual atomic properties in a molecule. Atom-specific topological features are integrated with various machine learning algorithms, including gradient boosting trees and convolutional neural network for protein thermal fluctuation analysis and B-factor prediction. Extensive numerical results indicate the proposed method provides a powerful topological tool for analyzing and predicting localized information in complex macromolecules.

Entities:  

Keywords:  Atom-specific topology; Convolutional neural network; Element-specific persistent homology; Gradient boosting tree; Protein flexibility

Year:  2020        PMID: 34278230      PMCID: PMC8281920          DOI: 10.1515/cmb-2020-0001

Source DB:  PubMed          Journal:  Comput Math Biophys        ISSN: 2544-7297


Introduction

In recent years tools from topology have been successfully applied to protein analysis [1, 2, 3, 4, 5, 6]. Topology offers one of highest level of abstractions of geometric data and allows one to infer high dimensional structure from low dimensional topological invariants. However, conventional topology oversimplifies geometry and thus lacks descriptive power for most real world problems. Persistent homology (PH) overcomes this difficulty by introducing a filtration parameter that describes the geometry in terms of a family of Betti numbers at various scales known as a barcode [7, 8, 9, 10]. Indeed, three dimensional (3D) protein spatial information from a protein data bank (PDB) file can be converted into a family of simplicial complexes. One can apply tools from algebraic topology to convert structural information into global topological invariants that provide a useful representation of biomolecular properties [11]. However, for quantitative biomolecular analysis and prediction, persistent homology alone neglects chemical and biology information. Element-specific persistent homology has been introduced to incorporate chemical and biological information into topological invariants [12, 13]. Similarity and differences between barcodes from different molecules can be measured by Wasserstein [14] and/or Bottleneck [15] distances. However, the previous applications of persistent homology and element-specific persistent homology are for the modeling and prediction of molecule-level thermodynamical or structural properties, such as protein-ligand binding affinities [13], protein folding free energy changes upon mutations [12, 16], drug toxicity [17], solubility, partition coefficient [18], and drug virtual screening (ligand and decoy classification) [19]. Essentially, topology is a global tool that examines the connectivity and relationship among many atoms in a neighborhood as a whole. High dimensional topological invariants, such as Betti 1 and Betti 2, describe the collective behavior of many atoms [20]. Therefore, it is not clear how to represent atomic level property, such as the B-factor of an atom, by persistent homology. In proteins, beta factor (B-factor) or (Debye-Waller factor is a measure of the attenuation of X-ray scattering caused by thermal motion. The amplitude of the thermal motion of an atom is theoretically proportional to its B-factor during the structure determination from X-ray diffraction data. It is well known that biomolecular flexibility provides an important link between its structure and function. In particular, it has been shown that intrinsic structural flexibility correlates to meaningful protein conformational variations, reactivity and enzymatic function [21]. As such, the accurate prediction of protein B-factor is essential to our understanding of protein structure, function and dynamics [22]. Early methods used to predict protein B-factor were derived from Hooke’s Law and are known as elastic mass-and-spring networks. In these models, alpha carbons (C) of biological macromolecules are treated as a mass and spring network and motions are predicted based on a harmonic potential. Given a protein, each C is represented as a node in the network and edges are weighted based on a potential function. Nodes are connected by an edge if they fall within a pre-defined euclidean cutoff distance. This captures the local covalent and non-covalent interactions between an individual atom and nearby atoms. One of the first mass- and-spring methods used for protein B-factor prediction is normal mode analysis (NMA). Like most B-factor prediction methods, NMA is independent of time and uses a Hamiltonian interaction matrix. Eigenvalues of the matrix system correspond to characteristic frequencies of the protein and these frequencies correlate with protein B-factors. Low-frequency modes correlate with cooperative motion and can be useful for hinge detection and domain motion. NMA has also been successfully implemented to understand the deformation of supramolecular complexes. [21, 23, 24, 25] Elastic network model (ENM) was introduced as a more efficient model that significantly reduces computational cost compared to NMA through the use of a simplified spring network [26]. A specific example is anisotropic network model (ANM) [27]. Gaussian network model (GNM) further reduces the computational cost by ignoring the anisotropic motion, rendering a more accurate method for protein C B-factor analysis [28, 29, 30]. All of the aforementioned methods depend on matrix diagonalization, which has the computational complexity of , where N is the number of matrix atoms involved in the analysis. Recently, flexibility and Rigidity Index (FRI) methods have been proposed as a geometric graph approach to further reduce the computational cost. FRI methods rely on constructing a distance matrix using radial basis functions to scale atom to atom distance non-linearly [31]. All versions of FRI produce a flexibility index, that correlates to the B-factor, for each C. Several versions of FRI have been developed. Among them, fast FRI (fFRI) is of in computational complexity [32]. FRI methods are also more accurate than all of the earlier algebraic graph-based methods. Additionally, anisotropic FRI (aFRI) provides high quality anisotropic motion analysis [32]. Moreover, using several radial basis functions with different parametrizations, the multiscale flexibility rigidity index (mFRI) can successfully capture multiscale atomic interactions [33]. More recently, the authors introduced a multiscale weighted colored graph (MWCG) model. The MWCG is another geometric graph theory model that has been shown to be the best B-factor prediction model to date. First, element-specific interaction subgraphs are constructed based on selected atomic interactions between certain element types. Atoms are represented as graph nodes and subgraphs are generated using pairs of atoms of certain elements (e. g., carbon, nitrogen, oxygen, sulfur). A centrality metric that uses radial basis functions is applied to pairwise interactions in each subgraph. By varying the parametrization of the radial basis functions the MWCG model can capture multiple protein interaction scales. MWCG is unique in its ability to utilize both element specific and multiscale interactions for improved B-factor prediction [34]. Most recently, MWCG is incorporated with machine learning algorithms for across-protein blind predictions of protein B-factors [35]. The objective of the present work is to extend the utility of persistent homology for atomic level property modeling and prediction. To this end, we introduce atom-specific persistent homology (ASPH) to create a local atomic representation of an atom using a global topological tool in a novel way. Specifically, ASPH constructs a pair of conjugated sets of point clouds or atoms centered around the atom of interest. The first set of a pair of conjugated sets of atoms for a given atom is selected by a local sphere of radius r around the atom of interest. The second set of atoms is defined by excluding the atom of interest in the first set. Conjugated simplicial complexes, conjugated chain groups, conjugated homology groups as well as conjugated persistence barcodes or diagrams are induced by an identical filtration. Conjugated persistence barcodes are compared with Bottleneck and Wasserstein metrics. The resulting distance provides a global topological representation of the localized atomic property, such as protein flexibility analysis and atomic-level protein B-factor information. Obviously, the proposed atom-specific topology can be applied to a wide variety of chemical and biological problems where atomic properties are measured, such as the chemical shifts of nuclear magnetic resonance (NMR), the B-factors of X-ray structure determination, and the shift and line broadening of other atomic spectroscopy. We focus on protein C B-factor prediction but the approach provided in this work is a general framework that can be used to predict B-factors of any atom in a protein. First, we use the generated atom-specific persistent homology features to fit B-factors within a given protein using linear least squares minimization. Note that this method does work for blind B-factor predictions across proteins. Additionally, the atom-specific persistent homology features are combined with other local and global protein features to construct machine learning models for the blind prediction of protein B-factors across different proteins. Moreover, image-like multiscale atom-specific persistent homology features are generated using an early technique [36]. These image like features, together with other features, are fed into convolutional neural networks (CNN). Training and validation are carried out using a large and diverse set of proteins from the protein data bank (PDB). We demonstrate that the proposed method offers some of the best results for blind B-factor predictions of a set of 364 proteins.

Methods and algorithms

Atom-specific persistent homology

Overview

Topology describes (continuous) objects in terms of topological invariants, i.e., Betti numbers. Betti-0, Betti-1, and Betti-2 which can be interpreted as connected components, rings, cavities, etc. Table 1 provides examples of the Betti numbers of a point, circle, sphere, and torus.
Table 1:

Topological invariants displayed as Betti numbers. Betti-0 represents the number of connected components, Betti-1 the number of tunnels or circles, and Betti-2 the number of cavities or voids. Two auxiliary rings are added to the torus to illustrate that its Betti-1=2.

ExamplePointCircleSphereTorus
Betti-01111
Betti-10102
Betti-20011
Given discrete data points, such as a point cloud or the set of atoms in a molecule, we use simplicial complexes to describe the topological relationship, or connectivity of the point cloud, to systematically identify topological invariants. First, a few simplicial complexes, as shown in Figure 1, are made up of vertices, edges, triangles, and tetrahedrons, denoted 0-simplex, 1-simplex, 2-simplex, and 3-simplex, respectively. Homology groups constructed from simplicial complexes give rise topological invariants. Given discrete dataset or a set of protein atoms, nontrivial topological information is generated by persistent homology. This introduces a filtration parameter to create a family of simplexes, which leads to a family of simplicial complexes, homology groups and associated topological invariants. By continuously varying the filtration parameter over an interval, the topological relationship among a given set of atoms is systematically reset, rendering a family of homology groups and corresponding topological invariants, which can be plotted as a persistence diagram, or a set of barcodes. Both persistence diagrams and barcodes record the birth and death (appearance and cessation) of Betti numbers during the filtration process. Many simplicial complex definitions, which determine the rules of the corresponding topological relationship, have been proposed. Specifically, Vietoris-Rips (VR) complex, Čech complex, and alpha complex are commonly used.
Figure 1:

From left to right an example of a 0-simplex, 1-simplex, 2-simplex, and 3-simplex.

Persistent homology allows the extraction of topological invariants that are embedded in the high dimensional data space of biomolecules. The resulting topological invariants over the filtration, i.e., persistence diagrams or persistence barcodes of different molecules can be compared using Bottleneck and Wasserstein distances. The goal of atom-specific persistent homology is to extract topological information of a given atom in a molecule. To embed local atomic information into a global topological description, we construct a pair of conjugated sets of point clouds, namely the original dataset and a datset excluding the atom of interest. The Bottleneck and Wasserstein distances between these two persistence diagrams reveal the desirable topological information of the given atom.

Simplex and simplicial complex

A (geometric) simplex is a generalization of a triangle or tetrahedron to arbitrary dimensions. A k-simplex is a convex hull of k + 1 vertices represented by a set of affinely independent points where with d ≥ k is the set of points, σ is the k-simplex, and constraints on λ,’s ensure the formation of a convex hull. An affinely independent combination of points can have at most k + 1 points in . For example a 1-simplex is a line segment, a 2-simplex a triangle, and a 3-simplex a tetrahedron. A subset of the k + 1 vertices of a k simplex with m + 1 vertices forms a convex hull in a lower dimension and is called an m-face of the k-simplex. An m-face is proper is m < k. The boundary of a k-simplex σ, is defined as the alternating sum of its (k + 1) faces, given as where [] denotes the convex hull formed by vertices of σ with the vertex u being excluded and ∂ is called the boundary operator. A collection of finitely many simplicies forms a simplicial complex denoted by . All simplicial complexes satisfy the following conditions. Faces of any simplex in are also simplices in . The intersection of any two simplicies σ1, is a face of both σ1 and σ2.

Homology

Given a simplicial complex , a k-chain c of is a formal sum of the k-simplices in and is defined as c = ∑ a where σ are the k-simplices and a’s coefficients. Generally, a are element of a field such as , , or . Computationally, it is common to choose a to be in . The group of k-chains in , denoted C, forms an Abelian group under addition in modulo two. This allows us to extend the definition of the boundary operator introduced in Eq. (2) to chains. The boundary operator applied to a k-chain c is defined as where σ’s are k-simplices. The boundary operator is a map from to , which is also known as a boundary map for chains. Note that in , the boundary operator ∂ satisfies the property that ∂ ‘ ∂ σ = 0 for any (k + 1)-simplex σ following the fact that any (k – 1)-face of σ is contained in exactly two k-faces of σ. The chain complex is defined as a sequence of chains connected by boundary maps with decreasing dimension and is denoted The k-cycle group and k-boundary group are then defined as kernel and image of ∂ and ∂ respectively, and where is the k-cycle group and is the k-boundary group. Since ∂ ‘ ∂ = 0, we have . Then the k-homology group is defined to be the quotient group of the k-cycle group modulo the k-boundary group, where is the k-homology group. The kth Betti number is defined to be rank of the k-homology group as .

Filtration and persistence

For a simplicial complex , we define a filtration of as a nested sequence of subcomplexes of , In persistent homology, the nested sequence of subcomplexes usually depends on a filtration parameter. The persistence of a topological feature is denoted graphically by its life span with respect to filtration parameter. Subcomplexes corresponding to various filtration parameters offer the topological fingerprints over multiple scales. The k persistence Betti number is given by the ranks of the k homology groups of that are alive at and are defined as The persistence of Betti numbers over the filtration interval can be recorded in many different ways. The commonly used ones are persistence barcodes and persistence diagrams. An example of barcodes is provided in Figure 2.
Figure 2:

(a) An example of 5 points in and (b) the corresponding persistence barcodes. The length of each barcode corresponds to the persistence of each topological object (β0, β1, β2, etc..) over the Vietoris-Rips (VR) complex filtration.

Similarity and distance

In this work, we use Bottleneck and Wasserstein distances to extract atom-specific topological information and facilitate atom-specific persistent homology. Let X and Y be multisets of data points, the Bottleneck and Wasserstein distances of X and Y are given by [15] and [14] respectively. Here B(X, Y) is the collection of all bijections from X to Y. Note that in our work, topological invariants of different dimensions are compared separately.

Vietoris-Rips complex

Given a metric space M and a cutoff distance d, a simplex is formed if all points have pairwise distances no greater than d. All such simplices form the Vietoris-Rips (VR) complex. The abstract nature of the VR complex allows the construction of simplicial complexes from a correlation function, which models the pairwise interaction of atoms using a radial basis function versus more standard distance metrics. The R library TDA is used to generate persistence barcodes [37].

Atom-specific persistent homology and element-specific persistent homology

Element-specific persistent homology was introduced to embed chemical and biology information into topological invariants [12, 19]. Its essential idea is to construct topological representations from subsets of atoms in various element types in a protein. For example, if one selects all carbon atoms in a protein, the resulting persistence barcodes will represent the strength and network of hydrophobicity in the protein. In contrast, atom-specific persistent homology is designed to highlight the topological information of a given atom in a biomolecule. It creates two conjugated subsets of atoms centered around the atom of interest, one with and one without the specific atom. Conjugated simplicial complexes, conjugated homology groups and conjugated topological invariants are generated for the conjugated sets of points clouds. The difference between the conjugated topological invariants, measured by both Wasserstein and Bottleneck distances, offers a topological representation of the atom of interest. As shown in Figure 3, atom-specific and element-specific conjugated point clouds can be constructed for a given dataset.
Figure 3:

Illustration of Atom-specific persistent homology point clouds. Top: the original point cloud. The atom of interest is at the center of the circle. Second row: a pair of conjugated sets of point clouds for atom-specific persistent homology. The rest: Four pairs of conjugated point clouds for atom-specific and element-specific persistent homology.

In this work, we focus on C B-factor predictions. We use element specific persistent homology to enhance the topological representation of each C neighborhood. Meanwhile, we develop atom-specific persistent homology to pinpoint the topological representation at each C atom. With these selections of subsets, Vietoris-Rips complexes are constructed by contact maps or matrix filtration [1]. To capture element-specific interactions we consider three subsets of carbon-carbon, carbon-nitrogen, and carbon-oxygen point clouds. This gives us the following element specific pairs, For a given Protein Data Bank (PDB) file, persistence barcodes are calculated as follows. Given a specific C of interest, say in an element specific set (, , and ), a point cloud consisting of all atoms within a pre-defined cutoff radius r is selected: where N is the number of atoms in the kth element pair . A conjugated set of point cloud, , includes the same set of atoms, except for . For a given pair of conjugated point clouds and , conjugated simplicial complexes, conjugated homology groups, and conjugated persistence barcodes are computed via persistent homology. We compute Euclidean distance based filtration using the Vietoris-Rips complex. Additionally, for a given set of atoms selected according to atom-specific and element specific constructions, we generate a family of multiresolution persistence barcodes by a resolution controlled filtration matrix: [1] where ϑ denotes a set of kernel parameters. We have used both exponential kernels and Lorentz kernels where η κ, and ν are pre-defined constants. This filtration matrix is used in association with the Vietoris-Rips complex to generate persistence barcodes or persistence diagrams. Then these topological invariants are compared using both Bottleneck and Wasserstein distances. An example of the conjugated persistence barcode pair generated for a C atom is illustrated in Figure 4.
Figure 4:

Illustration of atom-specific persistent homology using the fragments of protein 1AIE near residue 338 (i.e., residues 332-339). The left chart provides illustrations of the protein with and without C 338 from residue 338. The right chart shows conjugated persistence barcodes generated with and without C 338.

Machine learning models

Topological features are used for prediction of protein B-factor using both least squares fitting and machine learning as described in the following subsections.

Gradient boosted trees

Gradient boosting is an ensemble method that uses a number of “weak learners” to construct a prediction model in an iterative manner. The method is optimized via gradient descent, which minimizes the residuals of a loss function. At each step of the gradient boosting, gradient boosting trees (GBTs) incorporate decision trees to improve their predictive power. Ensemble methods like GBTs are useful because they can handle a diverse feature set, have strong predictive power, and are typically robust to outliers and against overfitting. In this work, we optimize the GBT hyper-parameters using the standard practice of a grid search. The parameters used for testing are provided in Table 2. Any hyper-parameters not listed in the table were taken to be the default values provided by the python scikit-learn package (version 0.21.3).
Table 2:

Boosted gradient tree hyperparameters used for testing. Parameters were determined using a grid search. Any hyperparameters that is not listed were taken to be the default values provided by the python scikit-learn package.

ParameterSetting
Loss FunctionQuantile
Alpha0.975
Estimators500
Learning Rate0.25
Max Depth4
Min Samples Leaf9
Min Samples Split9

Deep learning with a convolutional neural network

Neural networks are modeled after the function of neurons in brain. A neural network applies activation functions, called perceptrons, to inputs. Weights of the network are trained to minimize a loss function over many epochs, or passes of an entire training dataset. When a neural network has several layers of perceptrons we call it a deep neural network (DNN) and the intermediate layers are known as hidden layers. Convolutional neural networks (CNNs) have recently had great success in image classification. Using convolutions of a pre-defined filter size and number of filters, CNNs can automatically extract high-level features from input images. CNNs are advantageous because they can perform as well as other models without training as many parameters as a densely connected deep neural network. By applying several convolutions one can extract high-level features of an image. In this work we generate a image-like heat map by using a range of kernel parameters for atom-specific and element-specific persistent homology. The CNN output is then flattened and fed as input to a DNN along with global and local protein features. This allows us to use the same feature set as the boosted gradient method as well as the generated PH image data. A diagram of the CNN architecture is provided in Figure 5.
Figure 5:

The deep learning architecture using a convolutional neural network combined with a deep neural network. The plus symbol represents the concatenation of features.

For each C of the training set, the CNN is passed a three-channel persistent homology image of dimension (8,10,3). The model takes the input image data and applies two convolutional layers with 2x2 filters followed by a dropout of 0.5. The image data is passed through a dense layer, flattened, then joined with the other global and local features to form a dense layer of 218 neurons. This is followed by a dropout layer of 0.5, another dense layer of 100 neurons, a dropout layer of 0.25, a dense layer of 10 neurons, and finishes with a dense layer of output. Figure 5 provides an illustration of the deep CNN used in this work. The deep convolutional neural network has several hyper-parameters that can be tuned. As with the GBT, the deep convolutional neural network hyper-parameters are optimized using a basic grid search. Table 3 provides the parameters used for testing. Any hyper-parameters that are not listed below were taken to be the default values provided by the python Keras package.
Table 3:

Convolutional Neural Network (CNN) parameters used for testing. Parameters were determined using a grid search. Any hyper-parameters not listed below were taken to be the default values provided by python with the Keras package.

ParameterSetting
Learning Rate0.001
Epoch1000
Batch Size1000
LossMean Squared Error
OptimizerAdam

Consensus method

In this work, we combine the predictions of two machine learning models to construct a simple consensus model. The consensus prediction used in this work is generated by the average of C B-factor values predicted from the GBT and deep CNN models.

Machine learning features

A variety of element-specific and atom-specific persistence barcodes were generated using the techniques discussed in Sec. 2.1.7. In this work, we include 60 topological features. These features are generated in several ways by varying: kernels (Lorentz and exponential), element-specific pairs (CC, CN, CO), and distance metrics (Wasserstein-0 and Wasserstein-1, Bottleneck-0 and Bottleneck-1). For this work all persistent homology features were generated with the cutoff of 11Å.

Wasserstein and Bottleneck metrics for modified persistence diagrams

The distances evaluated from Wasserstein and Bottleneck evaluations of persistence diagrams depend on the boundary of the diagrams. Specifically, when two persistence diagrams are compared, the extra events on one diagram that do not match any events on the other diagram might contribute to the final distance by their distances from the boundary. For this reason, we create two additional persistence diagrams in which the y-axis is rotated clockwise by 30° or 60°, respectively, see Figure 6. This modification changes the Bottleneck and Wasserstein distances and allows the model to recognize elements that have a short persistence (i.e. have a short lifespan). Lastly, we modified the persistence diagram by reflecting around the diagonal axis. An example of this modification is illustrated in Figure 6. Table 4 provides a list of kernels, kernel parameters, y-axis change, distance metric, and element-specific pairs used to generate features in machine learning models.
Figure 6:

Illustration of modified persistence diagrams used in distance calculations. Left: Unchanged. Middle: Rotated 30°. Right: rotated 60°. Black dots are Betti-0 events and triangles are Betti-1 events.

Table 4:

Parameters used for topological feature generation. All features used a cutoff of 11Å. Both lorentz (Lor) and exponential (exp) kernels and Bottleneck (B) and Wasserstein (W) distance metrics were used.

No. featuresKernelKernel parameterDiagramDistance metricElement-specific pair
12Lorη = 21, ν = 5UnchangedB, WCC, CN, CO
12Expη = 10, κ = 1UnchangedB, WCC, CN, CO
12Expη = 2, κ = 1Diagonal reflectionB, WCC, CN, CO
12Expη = 2, κ = 1Rotated 30°B, WCC, CN, CO
12Expη = 2, κ = 1Rotated 60°B, WCC, CN, CO
Other features include global features from PDB files, i.e., R-value, protein resolution, and number of heavy atoms. Additional local features include packing density, amino acid type, occupancy, and secondary structure information generated by STRIDE software [38].

Image-like persistent homology features

Using the process described in Section 2.1.7 we generate 2D image-like persistent homology features, , for each C of the proteins in the dataset by using various values of η and κ in the kernel function. A cutoff of 11 Å with an exponential kernel and different values of η and κ are used to capture a wide variety of scales. In particular we use η = {1, 2, 3, 4, 5, 10, 15, 20}, and κ = {1, 2, 3, 4, 5, 6, 7, 8, 9, 10}. The image-like matrix is given by in Eq. (17), where each atom represents the PH feature of the i C atom, and k atom interaction (C, N, or O). This results in 2D PH images of dimension (8,10). Images are created for element-specific C interactions with carbon, nitrogen, and oxygen atom giving each image three channels. This results in a final image dimension of (8,10,3) for each C atom.

Results

Data sets

In this work, we use two data sets, one from Refs. [32, 33] and the other from Park, Jernigan, and Wu [39]. The first contains 364 proteins [32, 33] and the second contains 3 subsets of small, medium, and large proteins [39]. All sequences have a resolution of 3 Å or higher and an average resolution of 1.3 Å and the sets include proteins that range from 4 to 3912 residues [39]. For all testing, we exclude protein 1AGN due to known problems with this protein data [33]. Proteins 1NKO, 2OCT, and 3FVA are also excluded because these proteins have residues with B-factors reported as zero, which is unphysical. For the machine learning results, proteins 1OB4, 1OB7, 2OLX, and 3MD5 are excluded because the STRIDE software is unable to provide secondary features for these proteins. The image like features used in all convolutional neural networks were standardized with mean 0 and variance of 1

Evaluation metric

We use the proposed methods to predict B-factors of all C atoms present in a protein. Linear least square fitting was done using only topological features. The machine learning models were executed using a leave-one-(protein)-out method to blindly predict the B-factors of all C atoms in each protein. The machine learning models were trained using the data and features described in Sections 2.1.7, 2.2, 2.3. For comparison, we include previously existing C B-factor prediction fitting methods. To quantitatively assess our method for B-factor prediction we use the Pearson correlation coefficient given by where , i = 1, 2,…, N are predicted B-factors using the proposed method and , i = 1, 2,…, N experimental B-factors from the PDB file. The terms and represent the i theoretical and experimental B-factors respectively. Here and are averaged B-factors.

Cutoff distance

In this work, the optimal cutoff of r = 11Å is found over a grid search using various cutoff distances. Figure 7 displays the average Pearson correlation coefficient, obtained via fitting, over an entire dataset of 364 protein using all persistent homology metrics with various point cloud distance cutoffs.
Figure 7:

Average pearson correlation coefficient over the entire protein dataset fitting all 24 persistent homology features using various cuttoff distances.

For each protein we use the parameters listed in Table 5. The values used in this work were determined using the standard practice of a grid search.
Table 5:

Parameters used for the persistent homology element specific features with a cutoff of 11 Å.

Kernel Typeνηκ
Lorentz521-
Exponential-101

Least squares fitting within proteins

The Pearson correlation coefficients using least squares fitting for C B-factor prediction of small, medium, and large protein subsets are provided in Tables 12, 13, and 14 respectively. Results for the all proteins in the dataset are provided in Table 15. The average Pearson correlation coefficients for small, medium, large, and superset data sets are provided in Table 6. Table 6 includes fitting results using only Bottleneck, only Wasserstein, and using both Bottleneck and Wasserstein metrics. We also include results using only exponential kernel, only a Lorentz kernel, or both an exponential and Lorentz kernel for fitting. All results reported here PH features generated with a cutoff of 11Å and include three element-specific subsets (carbon-carbon, carbon-nitrogen, carbon-oxygen). Overall fitting methods using the various persistent homology features performed similarly. The best results came from using features generated by both exponential and Lorentz kernels and both Bottleneck and Wasserstein distances. Using both kernels and both distance metrics resulted in an average correlation coefficient of 0.73 for the superset.
Table 12:

Pearson correlation coefficients of least squares fitting C B-factor prediction of small proteins using 11Å cutoff. Two Bottleneck (B) and Wasserstein (W) metrics using various kernel choices are included.

B & WBW
PDB IDNExpLorBothExpLorBothExpLorBoth
1AIE310.970.880.990.780.640.900.900.770.96
1AKG160.820.661.000.600.530.720.530.560.87
1BX7510.860.740.890.790.680.820.810.690.82
1ETL121.001.001.000.680.871.000.950.981.00
1ETM121.001.001.000.450.740.860.700.831.00
1ETN121.001.001.000.960.920.990.700.921.00
1FF4650.770.720.800.700.650.750.680.680.76
1GK7390.950.940.980.910.930.950.880.920.94
1GVD560.750.680.840.670.630.690.610.620.66
1HJE131.001.001.000.720.791.000.670.571.00
1KYC150.960.991.000.920.930.990.880.881.00
1NOT131.001.001.000.820.861.000.860.811.00
1O06220.980.971.000.960.920.970.970.940.98
1P9I290.890.880.980.870.820.920.870.840.89
1PEF180.960.971.000.880.940.960.920.940.96
1PEN160.960.901.000.600.670.830.470.730.94
1Q9B440.790.760.940.580.590.690.690.570.71
1RJU360.810.740.910.750.690.810.620.650.72
1U06550.500.520.720.370.360.520.460.390.55
1UOY640.730.720.830.650.660.690.650.690.73
1USE470.660.750.910.500.520.720.460.530.64
1VRZ131.001.001.000.920.921.000.770.851.00
1XY281.001.001.000.990.951.000.910.911.00
1YJO61.001.001.001.001.001.001.001.001.00
1YZM460.870.900.950.820.720.880.860.840.90
2DSX520.540.500.780.370.300.560.410.360.55
2JKU380.890.750.950.850.650.880.830.600.88
2NLS360.750.660.880.610.320.760.490.470.69
2OL961.001.001.001.001.001.001.001.001.00
6RXN450.740.630.860.590.480.760.490.490.76
Table 13:

Pearson correlation coefficients of least squares fitting C B-factor prediction of medium proteins using 11Å cutoff. Two Bottleneck (B) and Wasserstein (W) metrics using various kernel choices are included.

B & WBW
PDB IDNExpLorBothExpLorBothExpLorBoth
1ABA870.670.670.760.540.620.680.560.630.70
1CYO880.710.690.780.660.580.680.650.590.67
1FK5930.530.590.710.490.500.580.490.500.55
1GXU890.750.780.820.720.610.750.690.720.77
1I71830.440.660.760.410.460.560.380.580.59
1LR7730.610.620.710.570.550.630.460.560.58
1N7E950.670.710.800.540.680.720.540.630.73
1NNX930.840.840.880.810.790.830.810.810.86
1NOA1130.630.650.720.600.570.630.530.570.59
1OPD850.350.290.570.260.210.360.290.190.36
1QAU1120.590.610.660.570.550.580.550.570.58
1R7J900.880.860.910.830.760.870.810.790.86
1UHA820.700.750.820.690.680.740.670.690.73
1ULR870.560.530.680.490.500.590.440.500.61
1USM770.620.610.810.570.530.660.610.580.65
1V05960.670.660.720.600.610.650.520.610.65
1W2L970.720.720.790.600.630.690.560.610.69
1X3O800.660.660.720.620.600.650.620.640.67
1Z21960.700.730.820.610.630.640.640.690.72
1ZVA750.850.850.940.840.780.920.830.810.86
2BF9350.940.730.970.700.650.780.890.710.92
2BRF1030.740.730.760.740.710.740.720.720.75
2CE01090.770.790.860.750.730.800.710.770.79
2E3H810.660.710.820.620.690.760.560.690.78
2EAQ890.810.770.860.790.720.810.770.760.82
2EHS750.750.730.810.720.710.740.690.710.73
2FQ3850.780.760.820.750.750.790.680.750.78
2IP6870.720.660.820.670.580.730.640.640.78
2MCM1120.800.800.850.780.770.810.750.770.82
2NUH1040.770.740.850.730.630.810.750.660.80
2PKT930.440.390.690.390.350.550.360.360.43
2PLT980.660.630.720.570.590.670.520.590.66
2QJL1070.450.520.630.420.460.500.410.490.51
2RB8930.810.780.840.780.750.800.740.760.81
3BZQ990.570.620.690.500.550.610.470.550.59
5CYT1030.530.520.650.490.460.540.430.480.50
Table 14:

Pearson correlation coefficients of least squares fitting C B-factor prediction of large proteins using 11Å cutoff. Two Bottleneck (B) and Wasserstein (W) metrics using various kernel choices are included.

B & WBW
PDB IDNExpLorBothExpLorBothExpLorBoth
1AHO660.750.780.880.720.730.790.530.650.75
1ATG2310.500.500.610.450.470.530.380.480.51
1BYI2380.500.510.580.410.460.490.440.480.54
1CCR1090.650.660.710.530.560.650.430.580.63
1E5K1880.670.680.740.660.670.680.630.670.69
1EW41060.580.600.730.520.510.550.550.550.62
1IFR1130.650.590.730.560.540.650.470.530.62
1NLS2380.810.780.860.750.650.830.800.720.82
1O082210.460.480.560.440.420.500.370.450.48
1PMY1230.710.700.760.620.590.670.680.690.71
1PZ41130.880.820.930.860.740.890.850.760.88
1QTO1220.590.590.650.480.460.530.550.520.56
1RRO1080.390.350.560.310.230.450.330.190.45
1UKU1020.800.810.840.780.800.800.740.800.80
1V701050.640.650.750.560.600.660.510.580.62
1WBE2060.530.470.630.430.380.550.360.420.48
1WHI1220.570.550.630.420.440.570.340.430.55
1WPA1070.700.690.790.610.520.710.660.560.70
2AGK2330.650.650.690.610.640.650.550.630.67
2C712250.450.380.560.290.330.420.230.300.48
2CG71100.320.440.630.290.310.360.300.330.41
2CWS2350.590.550.660.530.520.540.400.520.55
2HQK2320.800.790.830.700.740.800.680.760.81
2HYK2370.590.580.630.510.550.590.430.540.60
2I241130.470.440.690.400.400.480.450.400.49
2IMF2030.610.650.710.590.560.600.590.590.64
2PPN1220.570.610.740.510.590.630.440.570.63
2R161850.500.510.660.460.450.510.450.460.52
2V9V1490.600.510.660.530.480.560.550.500.62
2VIM1140.380.330.520.290.280.410.240.310.40
2VPA2170.730.750.780.720.710.730.680.730.74
2VYO2070.680.700.770.640.660.720.590.680.70
3SEB2380.630.660.770.620.610.680.610.620.67
3VUB1010.650.600.710.600.560.610.610.570.64
Table 15:

Pearson correlation coefficients of least squares fitting Cα B-factor prediction of all proteins using 11Å cutoff. Two Bottleneck (B) and Wasserstein (W) metrics using various kernel choices are included.

B & WBW
PDB IDNExpLorBothExpLorBothExpLorBoth
1ABA870.670.670.760.540.620.680.560.630.70
1AHO660.750.780.880.720.730.790.530.650.75
1AIE310.970.880.990.780.640.900.900.770.96
1AKG160.820.661.000.600.530.720.530.560.87
1ATG2310.500.500.610.450.470.530.380.480.51
1BGF1240.750.700.820.640.540.750.680.610.75
1BX7510.860.740.890.790.680.820.810.690.82
1BYI2380.500.510.580.410.460.490.440.480.54
1CCR1090.650.660.710.530.560.650.430.580.63
1CYO880.710.690.780.660.580.680.650.590.67
1DF4570.930.920.970.920.890.950.880.910.94
1E5K1880.670.680.740.660.670.680.630.670.69
1ES52600.580.570.650.510.550.580.440.560.60
1ETL121.001.001.000.680.871.000.950.981.00
1ETM121.001.001.000.450.740.860.700.831.00
1ETN121.001.001.000.960.920.990.700.921.00
1EW41060.580.600.730.520.510.550.550.550.62
1F8R19320.610.630.700.590.620.630.500.620.65
1FF4650.770.720.800.700.650.750.680.680.76
1FK5930.530.590.710.490.500.580.490.500.55
1GCO10440.630.640.660.590.630.630.530.630.65
1GK7390.950.940.980.910.930.950.880.920.94
1GVD560.750.680.840.670.630.690.610.620.66
1GXU890.750.780.820.720.610.750.690.720.77
1H6V29270.290.310.330.280.290.300.230.290.30
1HJE131.001.001.000.720.791.000.670.571.00
1I71830.440.660.760.410.460.560.380.580.59
1IDP4410.480.470.550.430.450.470.390.460.48
1IFR1130.650.590.730.560.540.650.470.530.62
1K8U870.720.740.850.670.640.710.650.670.75
1KMM14990.570.540.590.490.530.540.360.530.57
1KNG1440.520.510.610.430.470.510.430.500.53
1KR41070.570.480.600.390.470.530.450.450.54
1KYC150.960.991.000.920.930.990.880.881.00
1LR7730.610.620.710.570.550.630.460.560.58
1MF71940.560.590.670.550.570.590.500.580.59
1N7E950.670.710.800.540.680.720.540.630.73
1NKD590.730.690.890.560.580.630.550.650.75
1NLS2380.810.780.860.750.650.830.800.720.82
1NNX930.840.840.880.810.790.830.810.810.86
1NOA1130.630.650.720.600.570.630.530.570.59
1NOT131.001.001.000.820.861.000.860.811.00
1O06220.980.971.000.960.920.970.970.940.98
1O082210.460.480.560.440.420.500.370.450.48
1OB451.001.001.001.001.001.001.001.001.00
1OB751.001.001.001.001.001.001.001.001.00
1OPD850.350.290.570.250.210.360.290.190.36
1P9I290.890.880.980.870.820.920.870.840.89
1PEF180.960.971.000.880.940.960.920.940.96
1PEN160.960.901.000.600.670.830.470.730.94
1PMY1230.710.700.760.620.590.670.680.690.71
1PZ41130.880.820.930.860.740.890.850.760.88
1Q9B440.790.760.940.580.590.690.690.570.71
1QAU1120.590.610.660.570.550.580.550.570.58
1QKI39120.380.420.450.340.380.410.320.380.40
1QTO1220.590.590.650.480.460.530.550.520.56
1R291220.710.560.760.550.350.690.690.430.72
1R7J900.880.860.910.830.760.870.810.790.86
1RJU360.810.740.910.750.690.810.620.650.72
1RRO1080.390.350.560.310.230.450.330.190.45
1SAU1230.760.750.810.700.730.750.680.740.76
1TGR1110.770.760.830.720.700.740.740.730.75
1TZV1570.760.780.830.730.710.770.690.700.74
1U06550.500.520.720.370.360.520.460.390.55
1U7I2590.710.710.730.620.680.700.530.670.71
1U9C2200.660.650.740.610.570.640.610.600.67
1UHA820.700.750.820.690.680.740.670.690.73
1UKU1020.800.810.840.780.800.800.740.800.80
1ULR870.560.530.680.490.500.590.440.500.61
1UOY640.730.720.830.650.660.690.650.690.73
1USE470.660.750.910.500.520.720.460.530.64
1USM770.620.610.810.570.530.660.610.580.65
1UTG700.570.530.680.510.490.600.490.490.56
1V05960.670.660.720.600.610.650.520.610.65
1V701050.640.650.750.560.600.660.510.580.62
1VRZ131.001.001.000.920.921.000.770.851.00
1W2L970.720.720.790.600.630.690.560.610.69
1WBE2060.530.470.630.430.380.550.360.420.48
1WHI1220.570.550.630.420.440.570.340.430.55
1WLY3220.620.640.670.590.620.630.540.620.64
1WPA1070.700.690.790.610.520.710.660.560.70
1X3O800.660.660.720.620.600.650.620.640.67
1XY1160.970.961.000.730.660.870.810.890.99
1XY281.001.001.000.990.951.000.910.911.00
1Y6X860.560.530.620.500.490.590.500.520.56
1YJO61.001.001.001.001.001.001.001.001.00
1YZM460.870.900.950.820.720.880.860.840.90
1Z21960.700.730.820.610.630.640.640.690.72
1ZCE1390.840.830.880.830.770.850.810.780.82
1ZVA750.850.850.940.840.780.920.830.810.86
2A504690.640.630.700.540.600.670.410.580.67
2AGK2330.650.650.690.610.640.650.550.630.67
2AH19390.450.470.490.420.450.460.330.460.48
2B0A1910.590.600.690.500.580.620.480.590.63
2BCM4150.460.410.500.390.390.400.350.390.45
2BF9350.940.730.970.700.650.780.890.710.92
2BRF1030.740.730.760.740.710.740.720.720.75
2C712250.450.380.560.290.330.420.230.300.48
2CE01090.770.790.860.750.730.800.710.770.79
2CG71100.320.440.630.290.310.360.300.330.41
2COV5340.660.640.700.630.640.670.570.640.67
2CWS2350.590.550.660.530.520.540.400.520.55
2D5W12140.520.520.540.490.520.520.410.520.53
2DKO2530.750.720.790.720.690.750.680.690.72
2DPL5650.350.360.410.300.320.350.240.330.37
2DSX520.540.500.780.370.300.560.410.360.55
2E104390.600.590.650.510.580.610.430.570.62
2E3H810.660.710.820.620.690.760.560.690.78
2EAQ890.810.770.860.780.720.810.770.760.82
2EHP2460.630.650.710.580.620.650.520.620.64
2EHS750.750.730.810.720.710.740.690.710.73
2ERW530.620.410.840.330.260.600.310.280.49
2ETX3900.540.540.570.520.530.560.470.510.54
2FB61290.710.660.760.670.630.690.650.630.74
2FG11760.550.560.620.540.520.580.520.540.57
2FN95600.510.490.620.440.470.550.410.460.55
2FQ3850.780.760.820.750.750.790.680.750.78
2G69990.590.650.760.420.500.660.470.450.60
2G7O680.890.910.950.850.790.880.760.820.87
2G7S2060.630.600.660.590.580.630.540.590.63
2GKG1500.770.710.830.740.650.780.760.670.78
2GOM1210.470.520.640.420.420.450.440.470.53
2GXG1400.740.720.790.710.680.720.690.680.73
2GZQ2030.450.400.600.380.340.480.240.290.31
2HQK2320.800.790.830.700.740.800.680.760.81
2HYK2370.590.580.630.510.550.590.430.540.60
2I241130.470.440.690.400.400.480.450.400.49
2I493990.540.530.620.430.510.560.410.490.58
2IBL1080.690.710.750.660.670.700.650.680.71
2IGD610.670.720.840.610.640.740.610.660.74
2IMF2030.610.650.710.590.560.600.590.590.64
2IP6870.720.660.820.660.580.730.640.640.78
2IVY890.430.530.690.350.450.480.340.420.57
2J322440.770.720.850.730.680.770.730.680.77
2J9W2030.590.600.700.550.590.640.510.590.62
2JKU380.890.750.950.850.650.880.830.600.88
2JLI1120.870.810.900.820.700.850.850.780.86
2JLJ1210.780.750.800.710.650.740.740.710.76
2MCM1120.800.800.850.780.770.810.750.770.82
2NLS360.750.660.880.610.320.760.490.470.69
2NR71930.750.750.790.740.720.760.710.730.77
2NUH1040.770.740.850.730.630.810.750.660.80
2O6X3090.740.750.780.700.730.750.650.730.75
2OA21400.630.640.700.550.490.600.600.630.67
2OHW2570.350.390.480.290.320.350.270.340.38
2OKT3770.430.370.490.310.360.400.220.330.46
2OL961.001.001.001.001.001.001.001.001.00
2OLX41.001.001.001.001.001.001.001.001.00
2PKT930.440.390.690.400.350.550.360.360.43
2PLT980.660.630.720.570.590.670.520.590.66
2PMR830.690.680.800.590.620.680.650.650.69
2POF4280.620.560.660.480.550.600.440.540.63
2PPN1220.570.610.740.510.590.630.440.570.63
2PSF6080.430.450.530.410.440.450.370.420.44
2PTH1930.710.710.770.650.700.730.610.690.72
2Q4N12080.650.620.680.580.550.590.550.570.61
2Q5232960.650.660.700.620.560.640.630.570.65
2QJL1070.450.520.630.420.460.500.410.490.51
2R161850.500.510.660.460.450.510.450.460.52
2R6Q1490.710.720.760.660.680.700.620.650.67
2RB8930.810.780.840.780.750.800.740.760.81
2RE22490.640.650.700.570.590.610.590.600.63
2RFR1660.730.660.800.680.570.740.720.590.74
2V9V1490.600.510.660.530.480.560.550.500.62
2VE85150.460.480.550.420.410.440.400.430.47
2VH7940.590.540.680.520.490.630.420.490.54
2VIM1140.380.330.520.290.280.410.240.310.40
2VPA2170.730.750.780.720.710.730.680.730.74
2VQ41060.560.540.640.430.490.560.350.460.58
2VY81620.470.460.580.380.420.460.380.420.49
2VYO2070.680.700.770.640.660.720.590.680.70
2W1V5510.690.670.770.630.630.700.560.640.68
2W2A3500.600.590.650.570.560.590.540.570.60
2W6A1390.590.590.640.510.520.540.520.560.60
2WJ51100.630.550.790.590.520.680.590.530.64
2WUJ1030.690.680.790.620.520.650.670.590.71
2WW71610.440.480.600.400.420.500.330.430.49
2WWE1200.710.710.830.620.620.750.610.580.73
2X1Q2400.480.440.540.380.390.460.340.370.47
2X251670.620.610.730.560.570.640.570.570.64
2X3M1750.610.610.690.600.550.640.570.570.60
2X5Y1850.670.630.710.600.590.640.530.580.69
2X9Z2660.500.420.540.370.380.420.380.390.51
2XHF3100.620.620.670.580.560.600.550.620.63
2Y0T1110.690.680.830.600.610.680.560.640.70
2Y721830.710.710.780.690.690.720.660.700.71
2Y7L3230.680.700.720.660.680.690.580.690.69
2Y9F1490.750.720.780.650.690.710.580.700.74
2YLB4180.550.520.630.460.490.520.340.490.59
2YNY3260.630.670.750.600.620.630.560.630.66
2ZCM3480.420.390.490.340.350.400.240.320.43
2ZU13600.610.610.680.530.580.630.450.580.63
3A0M1460.740.760.840.680.700.720.610.730.78
3A7L1280.690.610.780.520.450.590.620.540.67
3AMC6140.540.530.640.470.500.540.370.510.57
3AUB1240.360.410.530.310.260.410.320.320.37
3B5O2490.550.580.660.520.560.630.460.550.57
3BA13120.670.660.720.640.650.680.600.650.70
3BED2620.610.550.670.530.530.560.440.530.61
3BQX1360.520.500.540.470.480.510.410.460.51
3BZQ990.570.620.690.500.550.610.470.550.59
3BZZ1030.600.630.680.510.580.610.450.500.59
3DRF5670.320.320.380.270.290.330.220.300.34
3DWV3590.670.630.690.620.620.660.540.620.65
3E5T2680.550.520.600.510.510.560.380.500.55
3E7R400.810.860.960.780.770.810.730.820.88
3EUR1500.490.460.530.390.430.470.310.420.47
3F2Z1480.760.780.840.750.760.780.690.770.78
3F7E2610.660.650.710.610.640.650.470.630.69
3FCN1850.600.650.750.560.590.640.540.590.67
3FE7890.690.650.760.580.600.670.540.630.70
3FKE2500.470.420.520.400.360.490.340.360.45
3FMY750.710.690.790.660.640.700.660.660.71
3FOD480.480.470.820.420.330.550.380.350.48
3FSO2380.820.820.850.770.740.770.770.810.82
3FTD2570.600.570.670.490.520.590.410.520.60
3G1S4180.440.510.680.410.450.510.380.450.49
3GBW1700.770.780.840.640.740.790.510.710.81
3GHJ1290.710.710.810.650.670.720.650.680.72
3HFO2160.750.720.820.700.630.750.650.690.74
3HHP13140.610.620.680.570.590.620.520.590.63
3HNY1700.590.560.640.470.520.570.420.490.56
3HP42010.600.610.720.570.540.640.430.560.62
3HWU1550.600.690.810.570.610.630.500.610.68
3HYD81.001.001.001.001.001.001.001.001.00
3HZ82000.580.590.660.550.530.560.520.540.58
3I2V1270.570.580.660.510.530.610.400.480.53
3I2Z1400.580.590.650.520.540.560.560.570.61
3I4O1540.630.640.730.580.590.600.560.630.66
3I7M1450.580.620.710.530.550.580.490.580.64
3IHS1730.620.670.740.580.540.600.580.600.62
3IVV1680.800.800.890.750.760.830.680.740.79
3K6Y2270.530.530.600.480.490.520.420.500.55
3KBE1660.620.610.650.570.600.620.520.600.61
3KGK1900.790.800.840.770.790.810.680.790.80
3KZD940.790.720.830.550.680.770.470.660.78
3L412190.610.620.710.590.600.660.570.590.67
3LAA1760.700.660.800.680.560.760.690.600.77
3LAX1180.810.810.860.800.760.830.770.780.82
3LG38460.400.380.410.360.370.400.320.370.41
3LJI2700.530.530.620.470.520.580.450.520.56
3M3P2440.470.440.690.400.400.580.250.350.48
3M8J1780.740.720.750.690.690.730.670.700.73
3M9J2500.570.560.590.530.540.560.390.530.56
3M9Q1900.530.520.590.500.510.530.460.500.51
3MAB1800.570.560.620.520.470.550.560.510.56
3MD4131.001.001.000.910.941.000.930.991.00
3MD5141.001.001.000.980.931.000.940.921.00
3MEA1700.580.580.680.570.570.640.480.570.59
3MGN2770.330.320.470.260.280.300.160.290.39
3MRE4460.400.380.450.320.360.400.240.350.41
3N113250.430.450.510.420.440.450.380.440.45
3NE02080.770.790.840.750.700.770.700.760.82
3NGG970.800.810.850.720.740.780.740.760.80
3NPV5000.440.440.500.400.420.440.360.430.47
3NVG61.001.001.001.001.001.001.001.001.00
3NZL700.680.610.840.530.490.660.590.550.67
3O0P1970.620.640.710.590.620.640.530.620.64
3O5P1470.640.600.710.550.570.600.530.560.64
3OBQ1500.590.590.660.460.490.580.530.560.58
3OQY2360.710.660.730.630.640.700.600.640.72
3P6J1450.750.730.810.690.710.730.610.710.75
3PD72160.650.660.720.620.600.650.600.610.65
3PES1660.700.720.790.580.630.700.520.600.66
3PID3870.500.490.560.440.480.530.370.460.51
3PIW1610.660.670.780.600.630.700.560.630.72
3PKV2290.500.520.630.430.480.530.350.500.57
3PSM940.830.780.880.790.770.830.680.760.79
3PTL2890.500.500.530.490.490.500.430.490.50
3PVE3630.450.450.590.370.390.440.410.420.45
3PZ93570.510.450.570.360.380.420.340.390.50
3PZZ121.001.001.000.950.901.000.940.801.00
3Q2X61.001.001.001.001.001.001.001.001.00
3Q6L1310.390.440.560.330.310.370.340.370.42
3QDS2840.630.620.690.590.590.650.510.590.64
3QPA2120.680.660.780.450.450.470.590.590.65
3R6D2220.650.660.730.620.630.650.530.640.69
3R871480.480.470.550.410.440.480.400.450.47
3RQ91650.510.470.610.410.440.520.390.450.56
3RY01280.440.450.540.400.400.470.410.420.47
3RZY1510.650.650.840.590.540.650.570.510.59
3S0A1320.390.430.520.330.340.380.320.310.37
3SD21000.650.670.770.640.630.690.560.630.67
3SEB2380.630.660.770.620.610.680.610.620.67
3SED1260.390.450.550.280.290.380.330.330.40
3SO61570.670.710.780.630.690.730.550.640.70
3SR36570.450.440.480.430.410.450.390.430.44
3SUK2540.530.540.640.460.480.540.470.490.57
3SZH7530.530.530.570.510.510.520.450.520.53
3T0H2090.760.730.780.720.690.740.680.710.76
3T3K1220.660.660.720.550.620.680.480.600.68
3T471450.540.540.780.450.450.620.430.470.54
3TDN3590.470.430.530.430.420.440.380.430.49
3TOW1550.660.650.740.580.610.660.530.600.65
3TUA2260.570.550.630.520.500.550.450.520.54
3TYS780.780.580.860.670.480.730.700.460.75
3U6G2760.440.390.540.390.370.450.270.350.48
3U97850.780.780.840.770.730.800.770.760.80
3UCI720.670.640.720.480.530.570.550.560.63
3UR86370.520.530.600.490.510.550.450.520.53
3US61590.600.560.670.550.490.620.530.460.59
3V1A590.740.570.950.510.530.770.390.460.68
3V752940.500.490.570.480.460.530.470.470.53
3VN01930.870.880.900.860.870.880.790.880.89
3VOR2190.640.580.700.560.520.630.530.550.63
3VUB1010.650.600.710.600.560.610.610.570.64
3VVV1120.640.640.790.550.480.650.570.490.58
3VZ91630.650.640.700.600.550.630.600.600.67
3W4Q8260.610.600.680.560.590.610.470.600.64
3ZBD2130.360.470.740.240.280.340.250.310.36
3ZIT1570.510.470.590.360.390.470.470.410.52
3ZRX2410.560.560.630.490.520.530.460.520.56
3ZSL1650.390.390.540.280.220.400.310.240.37
3ZZP740.400.300.470.190.270.310.120.220.40
3ZZY2260.650.670.690.630.630.640.590.630.64
4A021690.610.560.660.490.520.570.310.510.60
4ACJ1820.550.590.750.550.580.610.510.590.60
4AE71890.690.670.740.630.610.650.630.650.69
4AM13590.570.540.590.530.520.530.460.530.55
4ANN2100.500.480.570.420.430.480.360.420.47
4AVR1890.570.570.700.530.510.590.490.530.57
4AXY560.550.600.760.470.480.630.470.500.62
4B6G5590.700.710.750.670.690.720.600.690.73
4B9G2920.810.820.850.780.800.810.710.820.83
4DD54120.600.630.710.570.590.630.510.610.66
4DKN4230.590.580.630.520.540.560.420.550.61
4DND930.750.660.820.670.640.750.610.640.74
4DPZ1130.680.700.790.650.640.670.620.640.69
4DQ73380.450.460.510.370.440.490.290.400.46
4DT41700.760.740.780.700.680.720.700.700.73
4EK33130.580.630.650.550.560.580.530.590.60
4ERY3180.610.600.670.590.590.640.520.590.65
4ES1960.760.770.860.690.730.780.570.740.83
4EUG2250.610.610.670.540.600.620.510.580.62
4F014590.380.370.470.320.340.370.220.340.39
4F3J1430.570.630.660.520.590.610.470.580.60
4FR91450.650.620.780.630.580.700.580.570.64
4G1451.001.001.001.001.001.001.001.001.00
4G2E1550.750.640.850.590.610.740.680.610.80
4G5X5840.710.690.800.690.640.740.640.670.72
4G6C6760.430.440.500.400.440.460.240.430.45
4G7X2160.530.470.610.410.310.470.510.370.53
4GA21830.550.560.700.520.530.570.490.530.60
4GMQ940.730.770.840.680.660.720.670.630.72
4GS3900.650.680.740.600.640.680.510.660.70
4H4J2780.670.670.820.630.640.750.570.660.69
4H891750.390.500.670.330.370.390.350.400.42
4HDE1670.630.550.750.590.520.690.590.510.67
4HJP3080.620.610.650.570.550.590.580.580.62
4HWM1290.690.660.710.660.600.680.680.630.70
4IL7990.630.630.650.600.590.620.570.610.62
4J113770.660.630.680.620.610.630.630.610.66
4J5O2680.770.760.820.710.620.770.750.660.77
4J5Q1620.650.630.750.570.560.660.590.570.64
4J783050.480.480.560.430.440.500.380.470.53
4JG22020.630.630.740.610.610.640.580.600.63
4JVU2070.670.640.750.570.580.660.590.600.67
4JYP5500.590.600.690.520.570.610.380.580.61
4KEF1450.520.490.650.400.420.490.270.450.56
5CYT1030.530.520.650.490.460.540.430.480.50
6RXN450.740.630.860.590.480.760.490.490.76
Table 6:

Average Pearson correlation coefficients of least squares fitting C B-factor prediction of small, medium, large, and superset using 11Å cutoff. Two Bottleneck (B) and Wasserstein (W) metrics using various kernel choices are included. Results for pFRI are taken from Opron et al[32]. GNM and NMA value are taken from the course grained C results reported in Park et al [39].

H
B & WBWpFRIGNMNMA
ExpLorBothExpLorBothExpLorBoth
Small0.870.840.940.740.720.850.740.730.860.590.540.48
Medium0.680.680.780.620.610.690.600.630.690.610.550.48
Large0.610.600.700.540.540.610.510.550.620.590.530.49
Superset0.650.640.730.580.580.650.550.590.660.630.57NA

Blind machine learning prediction across proteins

The aforementioned least squares fitting methods cannot predict the B-factors of unknown proteins. Machine learning methods enable us to blindly predict B-factors across proteins. In this section, we utilize both boosted gradient and convolutional neural network algorithms for the blind prediction of B-factor across different proteins. Taken together, the entire dataset contains more than 620 000 atoms. We use a leave-one-protein out cross validation in our prediction. That is, for each protein, the data from a protein whose B-factors will be predicted, is excluded from the training data. This gives rise to a training set of roughly 600 000 data points for each protein (i.e., atoms and associated B-factors). The Pearson correlation coefficients using boosted gradient (GBT), convolutional neural network (CNN), and consensus method (CON) for C B-factor prediction of small, medium, and large protein subsets are provided in Tables 8, 9, and 10 respectively. Parameters for GBT and CNN methods can be found in Tables 2 and 3. The global and local features used for training and testing are provided in Section 2.3. Results for all proteins are provided in Table 11. The average Pearson correlation coefficients for small, medium, large, and superset data sets are provided in Table 7. All results reported here use a cutoff of 11Å and include three element-specific subsets (carbon-carbon, carbon-nitrogen, carbon-oxygen). Kernel parameters for both exponential and Lorentz kernels are provided in Table 5. Results from previously existing C B-factor prediction methods are included for comparison in Table 7. Overall both GBT and CNN algorithms perform similarly. As expected, the CNN method outperforms the GBT with average correlation coefficients over the superset of 0.60 and 0.59, respectively. The consensus method improves upon both results with an average Pearson correlation coefficient of 0.61 over the superset. Table 7 shows that the blind prediction machine learning models perform better than fitting models GNM and NMA and similar to the pFRI fitting model.
Table 8:

Pearson correlation coefficients for cross protein C atom blind B-factor prediction obtained by boosted gradient (GBT), convolutional neural network (CNN), and consensus (CON) for the small-sized protein set.

PDB IDNGBTCNNCON
1AIE310.750.70.78
1AKG160.270.320.29
1BX7510.740.740.76
1ETL120.370.820.55
1ETM120.370.630.43
1ETN120.070.480.13
1FF4650.610.660.64
1GK7390.770.90.82
1GVD560.710.550.69
1HJE130.840.750.9
1KYC150.620.690.66
1NOT130.690.960.8
1O06220.940.930.95
1P9I290.730.730.74
1PEF180.790.820.82
1PEN160.360.740.44
1Q9B440.590.850.67
1RJU360.60.460.58
1U06550.440.40.45
1UOY640.720.70.76
1USE470.050.320.12
1VRZ130.540.340.54
1XY280.790.820.81
1YJO60.7−0.060.57
1YZM460.690.640.7
2DSX520.340.340.36
2JKU380.570.710.66
2NLS360.230.470.29
2OL960.940.850.94
6RXN450.590.60.61
Table 9:

Pearson correlation coefficients for cross protein C atom blind B-factor prediction obtained by boosted gradient (GBT), convolutional neural network (CNN), and consensus (CON) for the medium-sized protein set.

PDB IDNGBTCNNCON
1ABA870.730.710.74
1CYO880.640.70.68
1FK5930.590.60.61
1GXU890.670.680.69
1I71830.530.580.56
1LR7730.620.610.64
1N7E950.630.580.65
1NNX930.780.790.8
1NOA1130.550.530.56
1OPD850.420.340.41
1QAU1120.510.590.57
1R7J900.710.770.75
1UHA820.710.740.73
1ULR870.540.530.56
1USM770.730.720.75
1V05960.60.640.63
1W2L970.430.50.47
1X3O800.410.430.44
1Z21960.680.650.69
1ZVA750.70.70.71
2BF9350.480.790.58
2BRF1030.720.770.75
2CE01090.60.660.64
2E3H810.650.680.67
2EAQ890.570.630.61
2EHS750.620.670.65
2FQ3850.770.820.81
2IP6870.60.660.63
2MCM1120.710.770.75
2NUH1040.720.560.7
2PKT930.01−0.04−0.01
2PLT980.520.530.54
2QJL1070.540.570.56
2RB8930.670.70.7
3BZQ990.450.530.49
5CYT1030.390.340.39
Table 10:

Pearson correlation coefficients for cross protein C atom blind B-factor prediction obtained boosted gradient (GBT), convolutional neural network (CNN), and consensus (CON) for the large-sized protein set.

PDB IDNGBTCNNCON
1AHO660.660.660.7
1ATG2310.550.510.55
1BYI2380.610.50.6
1CCR1090.550.60.59
1E5K1880.740.720.74
1EW41060.590.60.61
1IFR1130.70.640.7
1NLS2380.550.570.57
1O082210.490.470.49
1PMY1230.590.70.65
1PZ41130.720.80.77
1QTO1220.530.480.54
1RRO1080.40.450.43
1UKU1020.750.760.77
1V701050.630.620.64
1WBE2060.60.560.6
1WHI1220.590.560.6
1WPA1070.650.650.67
2AGK2330.670.630.67
2C712250.570.60.6
2CG71100.30.320.32
2CWS2350.610.470.6
2HQK2320.770.770.78
2HYK2370.650.630.65
2I241130.440.460.46
2IMF2030.530.580.56
2PPN1220.640.540.63
2R161850.440.490.46
2V9V1490.530.520.54
2VIM1140.440.470.47
2VPA2170.660.750.71
2VYO2070.60.630.63
3SEB2380.630.60.63
3VUB1010.590.550.59
Table 11:

Pearson correlation coefficients for cross protein C atom blind B-factor prediction obtained by boosted gradient (GBT), convolutional neural network (CNN), and consensus method (CON) for the Superset.

PDB IDNGBTCNNCONPDB IDNGBTCNNCON
1ABA870.730.710.742X5Y1850.760.680.76
1AHO660.660.660.72X9Z2660.490.520.52
1AIE310.750.70.782XHF3100.580.570.58
1AKG160.270.320.292Y0T1110.710.710.74
1ATG2310.550.510.552Y721830.650.710.69
1BGF1240.610.580.622Y7L3230.660.660.68
1BX7510.740.740.762Y9F1490.740.750.76
1BYI2380.610.50.62YLB4180.670.660.7
1CCR1090.550.60.592YNY3260.650.710.69
1CYO880.640.70.682ZCM3480.330.380.36
1DF4570.850.850.882ZU13600.660.660.68
1E5K1880.740.720.743A0M1460.530.60.59
1ES52600.650.620.663A7L1280.440.610.53
1ETL120.370.820.553AMC6140.680.640.69
1ETM120.370.630.433AUB1240.50.50.55
1ETN120.070.480.133B5O2490.490.550.52
1EW41060.590.60.613BA13120.620.590.63
1F8R19320.520.540.543BED2620.450.530.5
1FF4650.610.660.643BQX1360.560.550.58
1FK5930.590.60.613BZQ990.450.530.49
1GCO10440.470.470.53BZZ1030.380.510.44
1GK7390.770.90.823DRF5670.510.450.52
1GVD560.710.550.693DWV3590.630.550.63
1GXU890.670.680.693E5T2680.440.480.46
1H6V29270.260.340.343E7R400.720.660.77
1HJE130.840.750.93EUR1500.360.420.38
1I71830.530.580.563F2Z1480.730.760.75
1IDP4410.620.60.633F7E2610.650.690.68
1IFR1130.70.640.73FCN1850.630.650.66
1K8U870.570.60.593FE7890.520.550.54
1KMM14990.640.510.633FKE2500.510.510.54
1KNG1440.50.520.513FMY750.650.670.68
1KR41070.560.710.633FOD480.450.570.54
1KYC150.620.690.663FSO2380.720.750.74
1LR7730.620.610.643FTD2570.640.680.67
1MF71940.650.660.673G1S4180.60.570.61
1N7E950.630.580.653GBW1700.740.740.75
1NKD590.70.70.723GHJ1290.580.560.59
1NLS2380.550.570.573HFO2160.510.570.54
1NNX930.780.790.83HHP13140.610.650.65
1NOA1130.550.530.563HNY1700.610.60.62
1NOT130.690.960.83HP42010.560.580.58
1O06220.940.930.953HWU1550.580.650.62
1O082210.490.470.493HYD80.990.740.99
1OPD850.420.340.413HZ82000.450.540.48
1P9I290.730.730.743I2V1270.440.520.48
1PEF180.790.820.823I2Z1400.60.60.6
1PEN160.360.740.443I4O1540.620.720.66
1PMY1230.590.70.653I7M1450.440.570.49
1PZ41130.720.80.773IHS1730.610.620.64
1Q9B440.590.850.673IVV1680.830.820.84
1QAU1120.510.590.573K6Y2270.560.570.58
1QKI39120.340.450.383KBE1660.560.640.6
1QTO1220.530.480.543KGK1900.760.80.78
1R291220.560.590.593KZD940.550.670.6
1R7J900.710.770.753L412190.610.640.64
1RJU360.60.460.583LAA1760.350.490.42
1RRO1080.40.450.433LAX1180.740.690.74
1SAU1230.540.660.593LG38460.450.510.5
1TGR1110.660.690.693LJI2700.570.550.58
1TZV1570.740.770.763M3P2440.530.590.57
1U06550.440.40.453M8J1780.720.710.74
1U7I2590.710.740.743M9J2500.560.520.56
1U9C2200.570.590.593M9Q1900.40.480.45
1UHA820.710.740.733MAB1800.630.630.65
1UKU1020.750.760.773MD4130.880.960.96
1ULR870.540.530.563MEA1700.620.630.63
1UOY640.720.70.763MGN2770.080.090.09
1USE470.050.320.123MRE4460.540.540.57
1USM770.730.720.753N113250.510.470.52
1UTG700.620.640.663NE02080.670.730.71
1V05960.60.640.633NGG970.720.750.75
1V701050.630.620.643NPV5000.510.50.54
1VRZ130.540.340.543NVG60.510.630.71
1W2L970.430.50.473NZL700.560.580.57
1WBE2060.60.560.63O0P1970.680.720.71
1WHI1220.590.560.63O5P1470.60.590.61
1WLY3220.640.620.663OBQ1500.590.570.59
1WPA1070.650.650.673OQY2360.660.590.66
1X3O800.410.430.443P6J1450.660.720.69
1XY1160.820.750.833PD72160.680.70.71
1XY280.790.820.813PES1660.560.540.57
1Y6X860.50.460.513PID3870.480.30.45
1YJO60.7−0.060.573PIW1610.720.770.75
1YZM460.690.640.73PKV2290.520.510.53
1Z21960.680.650.693PSM940.80.770.82
1ZCE1390.70.740.733PTL2890.530.550.55
1ZVA750.70.70.713PVE3630.610.610.63
2A504690.60.540.63PZ93570.610.580.63
2AGK2330.670.630.673PZZ120.940.850.93
2AH19390.480.550.543Q2X60.950.720.93
2B0A1910.620.590.633Q6L1310.470.530.52
2BCM4150.50.510.523QDS2840.620.620.63
2BF9350.480.790.583QPA2120.550.670.59
2BRF1030.720.770.753R6D2220.650.740.69
2C712250.570.60.63R871480.470.450.48
2CE01090.60.660.643RQ91650.460.40.46
2CG71100.30.320.323RY01280.410.490.46
2COV5340.740.720.753RZY1510.650.620.66
2CWS2350.610.470.63S0A1320.530.490.54
2D5W12140.540.640.593SD21000.560.560.57
2DKO2530.780.780.83SEB2380.630.60.63
2DPL5650.410.360.423SED1260.530.520.55
2DSX520.340.340.363SO61570.650.650.66
2OCT4390.640.670.673SR36570.50.460.5
2E3H810.650.680.673SUK2540.580.590.6
2EAQ890.570.630.613SZH7530.690.670.71
2EHP2460.660.620.673T0H2090.710.70.73
2EHS750.620.670.653T3K1220.760.760.78
2ERW530.120.240.163T471450.510.620.57
2ETX3900.490.480.513TDN3590.470.490.49
2FB61290.730.750.753TOW1550.610.630.63
2FG11760.570.610.593TUA2260.620.560.63
2FN95600.570.540.583TYS780.660.740.72
2FQ3850.770.820.813U6G2760.530.460.52
2G69990.620.50.63U97850.670.720.71
2G7O680.720.860.83UCI720.420.420.43
2G7S2060.550.580.583UR86370.640.60.64
2GKG1500.560.640.593US61590.610.630.64
2GOM1210.690.590.693V1A590.570.270.55
2GXG1400.650.670.683V752940.490.560.53
2GZQ2030.340.40.373VN01930.850.850.86
2HQK2320.770.770.783VOR2190.470.480.48
2HYK2370.650.630.653VUB1010.590.550.59
2I241130.440.460.463VVV1120.560.570.57
2I493990.650.610.663VZ91630.720.640.72
2IBL1080.650.660.673W4Q8260.650.60.66
2IGD610.570.560.583ZBD2130.550.490.55
2IMF2030.530.580.563ZIT1570.520.420.5
2IP6870.60.660.633ZRX2410.540.60.58
2IVY890.510.450.513ZSL1650.490.570.53
2J322440.750.790.793ZZP740.380.480.42
2J9W2030.640.580.643ZZY2260.650.650.68
2JKU380.570.710.664A021690.590.650.62
2JLI1120.620.680.654ACJ1820.620.660.64
2JLJ1210.710.710.744AE71890.650.70.68
2MCM1120.710.770.754AM13590.540.520.55
2NLS360.230.470.294ANN2100.440.430.45
2NR71930.780.760.794AVR1890.560.530.56
2NUH1040.720.560.74AXY560.590.650.62
2O6X3090.760.760.784B6G5590.690.680.71
2OA21400.540.550.564B9G2920.740.740.76
2OHW2570.560.460.544DD54120.610.620.63
2OKT3770.420.420.434DKN4230.660.640.68
2OL960.940.850.944DND930.620.670.65
2PKT930.01−0.04−0.014DPZ1130.70.740.72
2PLT980.520.530.544DQ73380.550.60.57
2PMR830.60.630.634DT41700.670.690.69
2POF4280.620.60.664EK33130.60.580.61
2PPN1220.640.540.634ERY3180.570.590.59
2PSF6080.420.420.434ES1960.690.690.71
2PTH1930.690.70.714EUG2250.560.550.58
2Q4N12080.440.430.454F014590.350.260.33
2Q5232960.550.280.524F3J1430.580.630.62
2QJL1070.540.570.564FR91450.60.560.61
2R161850.440.490.464G145−0.280.450.04
2R6Q1490.630.620.654G2E1550.750.720.76
2RB8930.670.70.74G5X5840.710.730.74
2RE22490.650.660.684G6C6760.560.540.58
2RFR1660.610.690.664G7X2160.450.40.45
2V9V1490.530.520.544GA21830.610.530.61
2VE85150.550.550.584GMQ940.760.670.76
2VH7940.750.560.734GS3900.610.560.61
2VIM1140.440.470.474H4J2780.750.740.77
2VPA2170.660.750.714H891750.530.580.56
2VQ41060.70.750.724HDE1670.660.720.7
2VY81620.770.680.764HJP3080.680.60.67
2VYO2070.60.630.634HWM1290.540.60.57
2W1V5510.640.690.664IL7990.550.550.56
2W2A3500.590.60.614J113770.580.490.58
2W6A1390.710.690.724J5O2680.670.680.69
2WJ51100.450.530.484J5Q1620.720.740.74
2WUJ1030.350.540.454J783050.630.60.64
2WW71610.360.350.374JG22020.720.720.73
2WWE1200.490.550.534JVU2070.70.70.72
2X1Q2400.440.50.474JYP5500.590.670.65
2X251670.50.570.554KEF1450.480.530.51
2X3M1750.640.650.655CYT1030.390.340.39
6RXN450.590.60.61
Table 7:

Average Pearson correlation coefficients C B-factor predictions for small-, medium-, and large-sized protein sets along with the entire superset of the 364 protein dataset. Gradient boosted tree (GBT), convolutional neural network, and consensus (CON) results are obtained by leave-one-protein-out (blind). The results of parameter-free flexibility-rigidity index (pf-FRI), Gaussian network model (GNM) and normal mode analysis (NMA) were obtained via the least squares fitting of individual proteins.

CNNGBTCONpFRIGNMNMA
Small0.630.580.620.590.540.48
Medium0.600.580.610.610.550.48
Large0.580.590.580.590.530.49
Superset0.600.590.610.630.57NA

Conclusion

An essential component of the paradigm of protein dynamics is the correlation between protein flexibility and protein function. The shear complexity and large number of degrees of freedom make quantitative understanding of flexibility and function an inherently difficult problem. Several time-independent methods for predicting protein B-factors exist. Examples include NMA [24, 40, 25, 23], ENM [26], GNM [28, 29, 41], and FRI methods [31, 32, 33, 42]. None of the methods above are able to blindly predict protein B-factors of an unknown protein. We hypothesize that the intrinsic physics of proteins lie in a low-dimensional space embedded in a high-dimensional data space. Based on this hypothesis the authors previously introduced the graph theory based multiscale weighted colored graph (MWCG) [34, 35]. The authors showed that MWCG’s are able to successfully blindly predict cross-protein B-factors. In this work we explore this hypothesis further by creating a B-factor predictor using tools from algebraic topology. In order to construct localized topological representations for individual atoms from global topological tools, we propose atom-specific topology and atom-specific persistent homology. This approach creates two conjugated sets of atoms: the first set is centered around the given atom of interest while the other set is identical but excludes the atom of interest. Element-specific selections are further implemented to embed biological information into atom-specific persistent homology. The distance between the topological invariants generated from these conjugated sets of atoms is used to represent the atom of interest. Both Bottleneck and Wasserstein metrics are utilized to estimate the topological distances between conjugated barcodes. The Vietoris-Rips complex is employed for topological barcode generation. To test the proposed method we use over 300 proteins or more than 600,000 B-factors. Atom-specific persistent homology features are generated using several element-specific interactions, kernel choices, parametrizations, and barcode distance metrics. First we employ topological features to fit protein B-factors using linear least squares. Using topological features our fitting model outperformed previous fitting models with an average Pearson correlation coefficient of 0.73 over the superset of proteins. Next we considered using the topological features to blindly predict protein B-factors of C atoms. We generated two machine learning models, a gradient boosted tree (GBT) and deep convolutional neural network (CNN). Additionally we averaged the C prediction from the two models to generate a more robust consensus model. A variety of local and global features were included in addition to the generated topological features. Our blind prediction consensus model outperformed both GNM and NMA fitting models and produced results similar to those of the pFRI fitting model. To the authors’ knowledge, this work is the first time persistent homology has been used to predict the B-factor of atoms in proteins. This approach is novel because topology is a global property and on its own cannot be directly used to describe local atomic information. Our unique approach allows us to create a localized topological representation using a global mathematical tool. This approach enables us to account for multiple spatial interaction scales and element specific interactions. Our results demonstrate that this is an accurate and robust topological approach. Moreover, the results could easily be improved by including a larger dataset, fine tuning parameters, and exploring different machine learning algorithms. This method can be applied to a variety of interesting applications related to molecules and biomolecules. Examples include allosteric site detection, hinge detection, hot spot identification, chemical shift analysis, atomic spectroscopy interpretation, and prediction of protein folding stability changes upon mutation. More generally this method may be amenable to problems outside chemistry and biology such as network dynamics and social network centrality measure.
  29 in total

1.  Anisotropy of fluctuation dynamics of proteins with an elastic network model.

Authors:  A R Atilgan; S R Durell; R L Jernigan; M C Demirel; O Keskin; I Bahar
Journal:  Biophys J       Date:  2001-01       Impact factor: 4.033

2.  Large Amplitude Elastic Motions in Proteins from a Single-Parameter, Atomic Analysis.

Authors: 
Journal:  Phys Rev Lett       Date:  1996-08-26       Impact factor: 9.161

3.  STRIDE: a web server for secondary structure assignment from known atomic coordinates of proteins.

Authors:  Matthias Heinig; Dmitrij Frishman
Journal:  Nucleic Acids Res       Date:  2004-07-01       Impact factor: 16.971

4.  The energy landscapes and motions of proteins.

Authors:  H Frauenfelder; S G Sligar; P G Wolynes
Journal:  Science       Date:  1991-12-13       Impact factor: 47.728

5.  Direct evaluation of thermal fluctuations in proteins using a single-parameter harmonic potential.

Authors:  I Bahar; A R Atilgan; B Erman
Journal:  Fold Des       Date:  1997

6.  Protein normal-mode dynamics: trypsin inhibitor, crambin, ribonuclease and lysozyme.

Authors:  M Levitt; C Sander; P S Stern
Journal:  J Mol Biol       Date:  1985-02-05       Impact factor: 5.469

7.  Stochastic model for protein flexibility analysis.

Authors:  Kelin Xia; Guo-Wei Wei
Journal:  Phys Rev E Stat Nonlin Soft Matter Phys       Date:  2013-12-10

8.  Integration of element specific persistent homology and machine learning for protein-ligand binding affinity prediction.

Authors:  Zixuan Cang; Guo-Wei Wei
Journal:  Int J Numer Method Biomed Eng       Date:  2017-08-16       Impact factor: 2.747

9.  Coarse grained normal mode analysis vs. refined Gaussian Network Model for protein residue-level structural fluctuations.

Authors:  Jun-Koo Park; Robert Jernigan; Zhijun Wu
Journal:  Bull Math Biol       Date:  2013-01-08       Impact factor: 1.758

10.  Persistent homology analysis of ion aggregations and hydrogen-bonding networks.

Authors:  Kelin Xia
Journal:  Phys Chem Chem Phys       Date:  2018-05-16       Impact factor: 3.676

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.