Literature DB >> 33108737

An Information-Theory-Based Approach for Optimal Model Reduction of Biomolecules.

Marco Giulini1,2, Roberto Menichetti1,2, M Scott Shell3, Raffaello Potestio1,2.   

Abstract

In theoretical modeling of a physical system, a crucial step consists of the identification of those degrees of freedom that enable a synthetic yet informative representation of it. While in some cases this selection can be carried out on the basis of intuition and experience, straightforward discrimination of the important features from the negligible ones is difficult for many complex systems, most notably heteropolymers and large biomolecules. We here present a thermodynamics-based theoretical framework to gauge the effectiveness of a given simplified representation by measuring its information content. We employ this method to identify those reduced descriptions of proteins, in terms of a subset of their atoms, that retain the largest amount of information from the original model; we show that these highly informative representations share common features that are intrinsically related to the biological properties of the proteins under examination, thereby establishing a bridge between protein structure, energetics, and function.

Entities:  

Mesh:

Substances:

Year:  2020        PMID: 33108737      PMCID: PMC7659038          DOI: 10.1021/acs.jctc.0c00676

Source DB:  PubMed          Journal:  J Chem Theory Comput        ISSN: 1549-9618            Impact factor:   6.006


Introduction

The quantitative investigation of a physical system relies on the formulation of a model of it, that is, an abstract representation of its constituents and the interactions among them in terms of mathematical constructs. In the realization of the simplest model that entails all of the relevant features of the system under investigation, one of the most crucial aspects is the determination of its level of detail. The latter can vary depending on the properties and processes of interest: the quantum-mechanical nature of matter is explicitly incorporated in ab initio methods,[1] while effective classical interactions are commonly employed in the all-atom (AA) force fields used in AA molecular dynamics (MD) simulations.[2,3] Representations of a molecular system whose resolution level is lower than the atomistic one are commonly dubbed coarse-grained (CG) models:[4−8] in this case, the fundamental degrees of freedom, or effective interaction centroids, are representatives of groups of atoms, and the interactions among these CG sites are parametrized so as to reproduce equilibrium properties of the reference system. An important distinction should be made between reproducing a given property and describing it. For example, it is evident that the explicit incorporation of the electronic degrees of freedom in the model of a molecule is necessary to reproduce its vibrational spectrum with qualitative and quantitative accuracy; on the other hand, the latter can be measured and described from the knowledge of the nuclear coordinates alone, i.e., from the inspection of a subset of the system’s degrees of freedom. This is a general feature, in that the understanding of a complex system’s properties and behavior can typically be achieved in terms of a reduced set of variables: statistical mechanics provides some of the most recognizable examples of this, such as the description of systems composed of an Avogadro’s number of atoms or molecules in terms of a handful of thermodynamical parameters. In computer-aided studies, and particularly in the fields of computational biophysics and biochemistry, recent technological advancements—most notably massive parallelization,[9] GPU computing,[10] and tailor-made machines such as ANTON[11]—have extended the range of applicability of atomistic simulations to molecular complexes composed of millions of atoms;[12−14] even in the absence of such impressive resources, it is now common practice to perform microseconds-long simulations of relatively large systems, up to hundred thousands of atoms. However, a process of filtering, dimensionality reduction, or feature selection is required in order to distill the physically and biologically relevant information from the immense amount of data in which it is buried. The problem is thus to identify the most synthetic picture of the system that entails all and only its important properties: an optimal balance is sought between parsimony and informativeness. This objective can be pursued by making use of the language and techniques of bottom-up coarse-grained modeling;[5,15] in this context, in fact, one defines a mapping operatorM that performs a transformation from a high-resolution configuration r (i = 1, ..., n) of the system described in great detail to a simpler, coarser configuration R (I = 1, ..., N < n) at lower resolution:where n and N are the number of atoms in the system and the number of CG sites chosen, respectively. The linear coefficients c in eq are constant, positive, and subject to the normalization condition ∑c = 1 to preserve translational invariance. Furthermore, coefficients are generally taken to be specific to each site,[15] that is, an atom i taking part in the definition of CG site I will not be involved in the construction of another site J (c = 0 ∀ J ≠ I). Once the mappingM is chosen, the interactions among CG sites must be determined. In this respect, several methodologies have been devised in the past decades to parametrize such CG potentials.[4−8] Some approaches aim at reproducing as accurately as possible the exact effective potential obtained through the integration of the microscopic degrees of freedom of the system, that is, the multibody potential of mean force (MB-PMF); this is achieved in practice by tuning the CG interactions so as to reproduce specific low-resolution structural properties of the reference systems.[16,17] Recently, other methods have been proposed that target not only the structure but also the energetics.[18,19] In this work, we do not tackle the issue of parametrizing approximate CG potentials but rather focus on the consequences of the simplification of the system’s description even if the underlying physics is the same, i.e., configurations are sampled with the reference AA probability. In other words, we focus purely on the effect of projecting the AA conformational ensemble onto a CG configurational space using the mapping as a filter. Inevitably, in fact, a CG representation loses information about the high-resolution reference,[5,20] and the amount of information lost depends only on the number and selection of the retained degrees of freedom. In CG modeling, the mapping is commonly chosen on the basis of general and intuitive criteria: for example, it is rather natural to represent a protein in terms of one single centroid per amino acid (usually the choice falls on the α-carbon of the backbone).[21] However, it is by no means assured that a given representation that is natural and intuitive to the human eye is also the one that allows the CG model to retain the largest amount of information about the original higher-resolution system.[22,23] A quantitative criterion to assess how much detail is lost upon structural coarsening is thus needed in order to make a sensible choice. In the past few years, various methods have been developed that target the problem of the automated construction of a simplified representation of a protein at a resolution level lower than atomistic. In a pioneering work, Gohlke and Thorpe proposed to partition a protein into a few sizewise diverse blocks, distributing the amino acids among the different domains so as to minimize the degree of internal flexibility of the latter.[24] This picture of a protein subdivided into quasi-rigid domains, which has been further developed by several other authors,[25−31] is founded on the notion of a simplified model where groups of atoms are assigned to coarse-grained sites not according to their chemistry (e.g., one residue ↔ one site) but rather on the basis of the local properties of the specific molecule under examination. These partitioning methods, however, employ only structural information, in that they aim to minimize each block’s internal strain, while the energetics of the system is neglected. Some approaches systematically reduce the number of atoms in a system’s representation by grouping them according to graph-theoretical procedures. For example, the method reported by De Pablo and co-workers maps the static structure on a graph and hierarchically decimates it by clustering together the “leaves”;[32] alternative methods lump residues in effective sites on the basis of a spectral analysis of the graph Laplacian.[33] More recently, Li et al.[34] developed a graph neural network-based method to match a data set of manually annotated CG mappings. Alternatively, it was proposed to retain only those atoms that guarantee the set of new interactions to quantitatively reproduce the MB-PMF.[22,35] However, these methods are based on linearized elastic network models[36−41] that have the remarkable advantage of being exactly solvable, thus allowing a direct comparison between the CG potential and MB-PMF, but cannot be taken as significant representations of the system’s highly nonlinear interactions. It follows that all of these pioneering approaches rely either on purely geometrical/topological information obtained from a single static structure; on an ensemble of structures, neglecting energetics and thermodynamics; or on extremely simplified representations of both structure and interactions that do not guarantee general applicability to systems of great complexity. Here we tackle the issue of the automated, unsupervised construction of the most informative simplified representation of biological macromolecules in purely statistical-mechanical terms, that is, in the language that is most naturally employed to investigate such systems. Specifically, we search for the mapping operator that, for a given number of atoms retained from the original AA model, provides a description whose information content is as close as possible to that of the reference. In this context, then, the term “coarse-grained representation” should not be interpreted as a system with effective interactions whose scope is to reproduce a certain property, phenomenon, or behavior; rather, the representations we discuss here are simpler pictures of the reference system evolving according to the reference microscopic Hamiltonian but viewed in terms of fewer degrees of freedom. Our objective is thus the identification of the most informative simplified picture among those possible. To this end, we make use of the concept of mapping entropy, Smap,[17,42−44] a quantity that measures the quality of a CG representation in terms of the “distance” between probability distributions—the Boltzmann distribution of the reference AA system and the equivalent distribution when the AA probabilities are projected into the CG coordinate space. The mapping entropy is ignorant of the parametrization of the effective interactions of the simplified model: Smap effectively compares the reference system, described through all of its degrees of freedom, to the same system in which configurations are viewed through “coarse-graining lenses”. The difference between these two representations lies only in the resolution, not in the microscopic physics. Recently, the introduction of a mapping-entropy-related metric proved to be a powerful instrument for determining the optimal coarse-graining resolution level for a biological system.[44] Applied to a set of model proteins, this method was capable of identifying the number of sites that needed to be employed in the simplified CG picture to preserve the maximum amount of thermodynamic information about the microscopic reference. However, such analysis was carried out at a fixed CG resolution distribution, with a homogeneous placement of sites along the protein sequence. Moreover, calculations were performed using an exactly solvable yet very crude approximation to the system’s microscopic interactions, namely, a Gaussian network model. Motivated by these results, in the following we develop a computationally effective protocol that enables the approximate calculation of the mapping entropy for an arbitrarily complex system. We employ this novel scheme to explore the space of the system’s possible CG representations, varying the resolution level as well as the distribution, with the objective of identifying the ones featuring the lowest mapping entropy—that is, allowing for the smallest amount of information loss upon resolution reduction. The method is applied to three proteins of substantially different size, conformational variability, and biological activity. We show that the choice of retained degrees of freedom, guided by the objective of preserving the largest amount of information while reducing the complexity of the system, highlights biologically meaningful and a priori unknown structural features of the proteins under examination, whose identification would otherwise require computationally more intensive calculations or even wet lab experiments.

Results

In this section we report the main findings of our work. Specifically, (i) we outline the theoretical and computational framework that constitutes the basis for the calculation of the mapping entropy; (ii) we illustrate the biological systems on which we apply the method; and (iii) we describe the results of the mapping entropy minimization for these systems and the properties of the associated mappings.

Theory

The concept of mapping entropy as a measure of the loss of information inherently generated by performing a coarse-graining procedure on a system was first introduced by one of us in the framework of the relative entropy method[17] and subsequently expanded in refs (42−44). For the sake of brevity, we here omit the formal derivation connecting the relative entropy Srel and the mapping entropy as well as a discussion of the former. A brief summary of the relevant theoretical results presented in refs (17) and (42−44) is provided in Appendix A. In the following we restrict our analysis to the case of decimation mappings M in which a subset of N < n atoms of the original system are retained while the remaining ones are integrated out, so that In this case, as shown in Appendix A, the mapping entropy Smap reads as[42]where DKL(p(r) ∥ p̅(r)) is the Kullback–Leibler (KL) divergence[45] between p(r), the probability distribution of the high-resolution system, and p̅(r), the distribution obtained by observing the latter through “coarse-graining glasses”. Following the notation of ref (42), p̅(r) is defined aswhere p(R) is the probability of the CG macrostate R, given byin which β = 1/kBT, u(r) is the microscopic potential energy of the system, and Z = ∫dr e–β is its canonical partition function, while Ω1(R) is defined aswhich is the degeneracy of the macrostate—how many microstates map onto the CG configuration R. The calculation of Smap in eq thus amounts to determining the distance (in the KL sense) between two, although both microscopic, conceptually very different distributions. In contrast to p(r), eq shows that p̅(r) assigns the same probability to all configurations that map onto the same CG macrostate R; this probability is given by the average of the original probabilities of these microstates. Importantly, p̅(r) represents the high-resolution description of the system that would be accessible only starting from its low-resolution description—i.e., p(R). Grouping together configurations into a CG macrostate has the effect of flattening the detail of their original probabilistic weights. An attempt to revert the coarse-graining procedure and restore atomistic resolution by reintroducing the mapping operator M in p(R) can only result in microscopic configurations that are uniformly distributed within each macrostate. Because of the smearing of probabilities, the coarse-graining transformations constitute a semigroup.[46] This irreversible character highlights a fundamental consequence of coarse-graining strategies: a loss of information about the system. The definition based on the KL divergence (eq ) is useful for practical purposes. A more direct understanding of this information loss and how it is encoded in the mapping entropy, however, can be obtained by considering the nonideal configurational entropies of the original and CG representations,which respectively quantify the information contained in the associated probability distributions p(r) and p(R):[47] the higher the entropy, the more uniform the distribution, which we associate with a lower information content. By virtue of Gibbs’ inequality, from eq one has Smap ≥ 0. Furthermore, as shown in Appendix A,so that the entropy of the CG representation is always higher than that of the reference microscopic representation, implying that a loss of information occurs in decreasing the level of resolution.[42,44] Critically, the difference between the two information contents is precisely the mapping entropy. The information that is lost in the coarse-graining process as quantified by Smap depends only on the mapping operator M—in our case, on the choice of the retained sites. This paves the way for the possibility of assessing the quality of a CG mapping on the basis of the amount of information about the original system that it is able to retain, a qualitative advancement with respect to the more common a priori selection of CG representations.[21] Unfortunately, eqs and 9 do not allow—except for very simple microscopic models (see ref (44))—a straightforward computational estimate of Smap for a system arising from a choice of its CG mapping, as the observables to be averaged involve logarithms of high-dimensional probability distributions and ultimately configuration-dependent free energies. However, having introduced the loss of information per macrostate Smap(R), defined by the relation[42,44]in Appendix B we show that this problem can be overcome by further subdividing microscopic configurations that map to a given macrostate according to their potential energy. Let us define Pβ(U|R) as the conditional probability that a system thermalized at inverse temperature β has energy U provided that is in macrostate R:Then Smap(R) can be exactly rewritten as follows (see Appendix B):where ⟨U⟩β| is the average of the potential energy restricted to the CG macrostate R, given by This derivation enables a direct estimate of the mapping entropy Smap from configurations sampled according to the microscopic probability distribution p(r). For a given mapping, the histogram of these configurations with respect to CG coordinates R and energy U approximates the conditional probability Pβ(U|R) and, consequently, Smap(R) (see eq ); the total mapping entropy can thus be obtained as a weighted sum of the latter over all CG macrostates (eq ). The only remaining difficulty consists of obtaining accurate estimates of the exponential average in eq , which are prone to numerical errors. As is often done in these cases (see, e.g., free energy calculations through Jarzynski’s equality or the free energy perturbation method[48,49]), it is possible to rely on a cumulant expansion of eq , which when truncated at second order providesInserting eq into eq results in a total mapping entropy given by For a CG representation to exhibit a mapping entropy of exactly zero, it is required that all microstates r that map onto a given macrostate R = M(r) have the same energy in the reference system. Indeed, in this case one has Pβ(U|R) = δ(U – u̅) in eq , where u̅ is the potential energy common to all microstates within macrostate R, and consequently, Smap(R) = 0. Equation highlights that deviations from this condition result in a loss of information associated with a particular CG macrostate that is proportional to the variance of the potential energy of all the atomistic configurations that map to R. The overall mapping entropy is an average of these energy variances over all macrostates, each one weighted with the corresponding probability. In the numerical implementation we thus seek to identify those mappings that cluster together atomistic configurations having the same energy, or at least very close energies, in order to minimize the information loss arising from coarse-graining. With respect to eq , we further approximate Smap as its discretized counterpart S̃map (see Methods):where we identify Ncl discrete CG macrostates R, each of which contributes to S̃map with its own probability p(R), taken as the relative population of the cluster. We then employ an algorithmic procedure to estimate and efficiently minimize, over the possible mappings, a cost function (see eq in Methods)defined as an average of values of S̃map computed over different CG configuration sets, each of which is associated with a given number of conformational clusters Ncl. Finally, it is interesting to note that the mapping entropy in the form presented in eq appears in the dual-potential approach recently developed by Lebold and Noid.[18,19] In those works, the authors obtained an approximate CG energy function E(R) that is able to accurately reproduce the exact energetics of the low-resolution system—i.e., the average energy ⟨U ⟩β| in macrostate R (see eq ). This was achieved by minimizing the functionalwith respect to the force field parameters contained in E, where the average in eq is performed over the microscopic model. Expressing ⟨U ⟩β| as a function of r through the mapping M allows χ2[E] to be decomposed as[18,19]Minimizing χ2[E] on E(R) for a given mapping as in refs (18) and (19) is tantamount to minimizing the second term in eq with the objective of reducing the error introduced by approximating ⟨U ⟩β| through E(R). However, a comparison of eqs and 19 shows that Smap coincides, up to a multiplicative factor, with the first term of eq . Critically, the latter depends only on the mapping M and would be nonzero also in the case of an exact parametrization of E, i.e., for E(R) ≡ ⟨U⟩β|. The approach illustrated in the present work goes in a direction complementary to that of refs (18) and (19), as we concentrate on identifying those mappings that minimize the one contribution to χ2[E] that is due to, and depends only on, the CG representation M.

Biological Structures

It is worth stressing that the results of the previous section are completely general and independent of the specific features of the underlying system. Of course, characteristics of the input such as the force field quality, the simulation duration, the number of conformational basins explored, etc. will impact the outcome of the analysis, as is necessarily the case in any computer-aided investigation; nonetheless, the applicability of the method is not prevented or limited by these features or other system properties, e.g., the specific molecule under examination, its complexity, its size, or its underlying all-atom modeling. To illustrate the method in its generality, here we focus our attention on three proteins that we chose to constitute a small yet representative set of case studies. These molecules cover a size range spanning from ∼30 to ∼400 residues and a similarly broad spectrum of conformational variability and biological function, and they can be taken as examples of several classes of enzymatic and non-enzymatic proteins. Each protein is simulated for 200 ns in the NVT ensemble with physiological ion concentration. Out of 200 ns, snapshots are extracted from each trajectory every 20 ps, for a total of 104 AA configurations per protein employed throughout the analysis. Details about the simulation parameters, quantitative inspection of MD trajectories, characteristic features of each protein’s results, and the validation of the latter with respect to the duration of the MD trajectory employed can be found in the Supporting Information. Hereafter we provide a description of each molecule along with a brief summary of its behavior as observed along MD simulations. The first protein is a recently released[50] 31-residue tamapin mutant (TAM, PDB code 6D93). Tamapin is the toxin produced by the Indian red scorpion. It features a remarkable selectivity toward a peculiar calcium-activated potassium channel (SK2), whose potential use in the pharmaceutical context has made it a preferred object of study during the past decade.[51,52] Throughout our simulation, almost every residue is highly solvent-exposed. Side chains fluctuate substantially, thus giving rise to extreme structural variability. The second protein is adenylate kinase (AKE, PDB code 4AKE), which is a 214 residue phosphotransferase enzyme that catalyzes the interconversion between adenine diphosphate (ADP) and adenine monophosphate (AMP) on the one hand, and the energy-rich complex adenine triphosphate (ATP) on the other hand.[53] It can be subdivided in three structural domains: CORE, LID, and NMP.[54] The CORE domain is stable, while the other two undergo large conformational changes.[55] Its central biochemical role in the regulation of the energy balance of the cell and its relatively small size, combined with the possibility to observe conformational transitions over time scales easily accessible by plain MD,[56] make it an ideal candidate to test and validate novel computational methods.[22,57,58] In our MD simulation, the protein displays many rearrangements in the two motile domains, which happen to be quite close at many points. Nevertheless, the protein does not undergo a full open ⇄ closed conformational transition. The third protein is α-1-antitrypsin (AAT, PDB code 1QLP). With 5934 atoms (372 residues), this protein is almost 2 times bigger than AKE. AAT is a globular biomolecule, and it is well-known to exhibit a conformational rearrangement over the time scales of minutes.[59−61] During our simulated trajectory, the molecule experiences fluctuations particularly localized in regions corresponding to the most solvent-exposed residues. The protein bulk appears to be very rigid, and there is no sign of a conformational rearrangement.

Minimization of the Mapping Entropy and Characterization of the Solution Space

The algorithmic procedure described in Methods and Appendix B enables one to quantify the information loss experienced by a system as a consequence of a specific decimation of its degrees of freedom. This quantification, which is achieved through the approximate calculation of the associated mapping entropy, opens the possibility of minimizing such a measure in the space of CG representations in order to identify the mapping that, for a given number of CG sites N, is able to preserve as much information as possible about the AA reference. In the following we allow CG sites to be located only on heavy atoms, thus reducing the maximum number of possible sites to Nheavy. We then investigate the properties of various kinds of CG mappings having different numbers of retained sites N. Specifically, we consider three chemically intuitive values of N for each biomolecule: (i) Nα, the number of Cα atoms in the structure (equal to the number of amino acids); (ii) Nαβ, the number of Cα and Cβ atoms; and (iii) Nbkb, the number of heavy atoms belonging to the main chain of the protein. The values of N for mappings (i)–(iii) in the cases of TAM, AKE, and AAT are listed in Table together with the corresponding values of Nheavy.
Table 1

Values of Nα, Nαβ, Nbkb, and Nheavy (See the Text) for Each Analyzed Protein

proteinNαNαβNbkbNheavy
TAM3159124230
AKE2144088561656
AAT37272314882956
Even with N restricted to Nα, Nαβ, and Nbkb, the combinatorial dependence of the number of possible decimation mappings on the number of retained sites and Nheavy makes their exhaustive exploration unfeasible in practice (see Methods). To identify the CG representations that minimize the information loss, we thus rely on a Monte Carlo simulated annealing (SA) approach (Methods).[62,63] For each analyzed protein and value of N, we perform 48 independent optimization runs, i.e., minimizations of the mapping entropy with respect to the CG site selection; we then store the CG representation characterized by the lowest value of Σ in each run, thus resulting in a pool of optimized solutions. In order to assess their statistical significance and properties, we also generate a set of random mappings and calculate the associated Σ values, which constitute our reference values. Figure displays, for each value of N considered, the distributions of mapping entropies obtained from random choices of the CG representations of TAM, AKE, and AAT together with each protein’s optimized counterpart. For N = Nbkb and N = Nα in Figure we also report the values of Σ associated with physically intuitive choices of the CG mapping that are commonly employed in the literature: the backbone mapping (N = Nbkb), which neglects all atoms belonging to the side chains; and the Cα mapping (N = Nα), in which we retain only the Cα atoms of the structures. The first is representative of united-atom CG models, while the second is a ubiquitous and rather intuitive choice to represent a protein in terms of a single bead per amino acid.[21]
Figure 1

Distributions of the values of the mapping entropy Σ [in kJ mol–1 K–1] in eq for random mappings (light-blue histograms) and optimized solutions (green histograms). Dark-blue dashed lines show the best fit with normal distributions over the random cases. Each column corresponds to an analyzed protein and each row to a given number N of retained atoms. In the first and last rows, corresponding to numbers of CG sites equal to the numbers of Cα atoms and backbone atoms (Nα and Nbkb, respectively), the values of the mapping entropy associated with the physically intuitive choice of the CG sites (see the text) are indicated by vertical lines (red for N = Nα, purple for N = Nbkb). It should be noted that the σ ranges have the same width in all of the plots.

Distributions of the values of the mapping entropy Σ [in kJ mol–1 K–1] in eq for random mappings (light-blue histograms) and optimized solutions (green histograms). Dark-blue dashed lines show the best fit with normal distributions over the random cases. Each column corresponds to an analyzed protein and each row to a given number N of retained atoms. In the first and last rows, corresponding to numbers of CG sites equal to the numbers of Cα atoms and backbone atoms (Nα and Nbkb, respectively), the values of the mapping entropy associated with the physically intuitive choice of the CG sites (see the text) are indicated by vertical lines (red for N = Nα, purple for N = Nbkb). It should be noted that the σ ranges have the same width in all of the plots. The optimality of a given mapping with respect to a random choice of the CG sites can be quantified in terms of the Z score,where μ and σ represent the mean and standard deviation, respectively, of the distribution of Σ over randomly sampled mappings. Table summarizes the values of Z found for each N for the proteins under examination, including Z[backbone] and Z[Cα], which were computed with respect to the random distributions generated with N = Nbkb and N = Nα respectively.
Table 2

Z Scores for Each Analyzed Proteina

 TAMAKEAAT
[Nα]–2.22 ± 0.06–7.85 ± 1.14–6.96 ± 1.03
Z[Nαβ]–2.38 ± 0.08–6.09 ± 0.79–6.64 ± 0.84
[Nbkb]–2.65 ± 0.09–5.55 ± 0.62–7.24 ± 0.85
Z[backbone]4.375.654.31
Z[Cα]0.873.363.28

We report the means (Z̅) and standard deviations of the distributions of Z values of the optimized solutions for all values of N investigated. Results for the standard mappings (Z[backbone] for backbone atoms only and Z[Cα] for Cα atoms only) are also included.

We report the means (Z̅) and standard deviations of the distributions of Z values of the optimized solutions for all values of N investigated. Results for the standard mappings (Z[backbone] for backbone atoms only and Z[Cα] for Cα atoms only) are also included. For the physically intuitive CG representations, Figure shows that the value of Σ associated with the backbone mapping is very high for all structures. For TAM in particular, the amount of information retained is so low that the mapping entropy is 4.37 standard deviations higher than the average of the reference distribution of random mappings (see Table ). This suggests that neglecting the side chains in a CG representation of a protein is detrimental, at least as far as the structural resolution is concerned. In fact, the backbone of the protein undergoes relatively minor structural rearrangements when exploring the neighborhood of the native conformation, thereby inducing negligible energetic fluctuations; for side chains, on the other hand, the opposite is true, with comparatively larger structural variability and a similarly broader energy range associated with it. Removing side chains from the mapping induces the clustering of atomistically different structures with different energies onto the same coarse-grained configuration, the latter being solely determined by the backbone. The corresponding mapping entropy is thus large—worse than a random choice of the retained atoms—since it is related to the variance of the energy in the macrostate. Calculations employing the Cα mapping for the three structures show that this provides Σ values that are very close to the ones we find with the backbone mapping, thus suggesting that Cα atoms retain about the same amount of information that is encoded in the backbone. This is reasonable given the rather limited conformational variability of the atoms along the peptide chain. However, a comparison of the random case distributions for Nα and Nbkb as the number of retained atoms in Figure reveals that the former generally has a broader spread than the latter because of the lower number of CG sites; consequently, the value of Σ for the Cα atom mapping is closer to the bulk of the distribution of the random case than that of the backbone mapping. We now discuss the case of optimized mappings, that is, CG representations retaining the maximum amount of information about the AA reference. Each of the 48 minimization runs, which were carried out for each protein in the set and value of N considered, provided an optimal solution—a deep local minimum in the space of CG mappings; the corresponding Σ values are spread over a compact range of values that are systematically lower than, and do not overlap with, those of the random case distributions (Figure ). The optimal solutions for AKE and AAT span wide intervals of Σ values; for N = Nα in particular, the support of this set and of the corresponding random reference have comparable sizes. A quantitative measure of this broadness is displayed in the distributions of Z scores for the optimal solutions presented in Table . In both proteins, we observe that the Σ values associated with the optimal mappings increase with the degree of coarse-graining, N; this is a consequence of keeping the number of CG configurations of each system (conformational clusters; see section ) constant across different resolutions. As N increases, the available CG conformational clusters are populated by more energetically diverse conformations, thereby incrementing the associated energy fluctuations. On the other hand, TAM shows narrowly peaked distributions of optimal values of Σ whose position does not vary with the number of retained sites. Both effects can be ascribed to the fact that most of the energy fluctuations in TAM—and consequently the mapping entropy—are due to a subset of atoms that are almost always maintained in each optimal mapping (see section ), in contrast to a random choice of the CG representation. At the same time, the associated Z scores are lower than the ones for the bigger proteins for all values of N under examination, as TAM conformations generally feature a lower variability in energy than the other molecules. For all of the investigated proteins, the absence of an overlap between the distributions of Σ associated with the random and optimized mappings raises some relevant questions. First, one might wonder what kind of structure the solution space has, that is, whether the identified solutions lie at the bottom of a rather flat vessel or, on the contrary, each of them is located in a narrow well, neatly separated from the others. Second, it is reasonable to ask whether some degree of similarity exists between these quasi-degenerate solutions of the optimization problem and, if so, what significance this has. In order to answer these questions, for each structure we select four pairs of mapping operators Mopt that result in the lowest values of Σ. We then perform 100 independent transitions between these solutions, constructing intermediate mappings by randomly swapping two non-overlapping atoms from the two solutions at each step and calculating the associated mapping entropies. Figure shows the results of this analysis for the pair of mappings with the lowest Σ; all of the other transitions are reported in Figure S2. It is interesting to notice that the end points (that is, the optimized mappings) correspond to the lowest values of Σ along each transition path; as the size of the protein increases, the values of Σ for intermediate mappings get closer to the average of Σrandom. By this analysis we cannot rule out the absence of lower minima over all of the possible paths, although it seems quite unlikely given the available sampling.
Figure 2

Values of the mapping entropy Σ [in kJ mol–1 K–1] for mappings connecting two optimal solutions. In each plot, one per protein under examination, the two lowest-Σ mappings are taken as initial and final end points (black dots) for paths constructed by swapping pairs of atoms between them (blue dots). For each protein, 100 independent paths at the given N = Nαβ were constructed, and the mapping entropy of each intermediate point was computed. In each plot, horizontal lines represent the mean (red) and minimum (green) values of Smap obtained from the corresponding distributions of random mappings presented in Figure .

Values of the mapping entropy Σ [in kJ mol–1 K–1] for mappings connecting two optimal solutions. In each plot, one per protein under examination, the two lowest-Σ mappings are taken as initial and final end points (black dots) for paths constructed by swapping pairs of atoms between them (blue dots). For each protein, 100 independent paths at the given N = Nαβ were constructed, and the mapping entropy of each intermediate point was computed. In each plot, horizontal lines represent the mean (red) and minimum (green) values of Smap obtained from the corresponding distributions of random mappings presented in Figure . Finally, it is interesting to observe the pairwise correlations of the site conservation probability within a pool of solutions, as it is informative about the existence of atom pairs that are, in general, simultaneously present, simultaneously absent, or mutually exclusive. As reported in detail in Figures S6 and S7, no clear evidence is available that conserving a given atom can increase or decrease in a statistically relevant manner the conservation probability of another: this behavior supports the idea that the organization internal to a given optimal mapping is determined in a nontrivial manner by the intrinsically multibody nature of the problem at hand. These analyses thus address the first question by showing that at least the deepest solutions of the optimization procedure are distinct from each other. It is not possible to (quasi)continuously transform an optimal mapping into another through a series of steps while keeping the value of the mapping entropy low. Each of the inspected solutions is a small town surrounded by high mountains in each direction, isolated from the others with no valleys connecting them. The second question, namely, what similarity (if any) exists among these disconnected solutions, is tackled in the following section.

Biological Significance

The degree of similarity between the optimal mappings can be assessed by a simple average, returning the frequency with which a given atom is retained in the 48 solutions of the optimization problem. Figure shows the values of Pcons, the probability of conserving each heavy atom, for each analyzed protein and degree of coarse-graining N investigated, computed as the fraction of times it appears in the corresponding pool of optimized solutions. One can notice the presence of regions that appear to be more or less conserved. Quantitative differences between the three cases under examination can be observed: while the heat map of TAM shows narrow and pronounced peaks of conservation probability, the optimal solutions for AKE feature a more uniform distribution, where the maxima and minima of Pcons extend over secondary structure fragments rather than small sets of atoms. The distribution gets even more blurred for AAT.
Figure 3

Probability Pcons that a given atom is retained in the optimal mapping at various numbers N of CG sites and for each analyzed protein, expressed as a function of the atom index. Atoms are ordered according to their numbers in the PDB file. The secondary structure of the proteins is depicted using Biotite:[64] green waves represent α-helices, and orange arrows correspond to β-strands.

Probability Pcons that a given atom is retained in the optimal mapping at various numbers N of CG sites and for each analyzed protein, expressed as a function of the atom index. Atoms are ordered according to their numbers in the PDB file. The secondary structure of the proteins is depicted using Biotite:[64] green waves represent α-helices, and orange arrows correspond to β-strands. As index proximity does not imply spatial proximity in a protein structure, we mapped the aforementioned probabilities to the three-dimensional configurations. Results for TAM are shown in Figure , while the corresponding ones for AKE and AAT are provided in Figure S3. From the distributions of Pcons for different numbers of retained sites N it is possible to infer some relevant properties of optimal mappings.
Figure 4

Structure of tamapin (one bead per atom) colored according to Pcons, the probability for each atom to be retained in the pool of optimal mappings. Each structure corresponds to a different number N of retained CG sites. Residues presenting the highest retainment probability across N (ARG6 and ARG13) are highlighted.

Structure of tamapin (one bead per atom) colored according to Pcons, the probability for each atom to be retained in the pool of optimal mappings. Each structure corresponds to a different number N of retained CG sites. Residues presenting the highest retainment probability across N (ARG6 and ARG13) are highlighted. With regard to TAM (Figure ), it seems that at the highest degree of CG (N = Nα), only two sites are always conserved, namely, two nitrogen atoms belonging to the ARG6 and ARG13 residues (Pcons(NH1, ARG6) = 0.92; Pcons(NH2, ARG13) = 0.96). The atoms that constitute the only other arginine residue, ARG7, are well-conserved but with lower probability. By increasing the resolution, i.e., employing more CG sites (N = Nαβ), we see that the atoms in the side chain of LYS27 appear to be retained more than average together with atoms of GLU24 (Pcons(NZ, LYS27) = 0.65; Pcons(OE2, GLU24) = 0.75). At N = 124, the distribution becomes more uniform but is still sharply peaked around terminal atoms of ARG6 and ARG13. Interestingly, ARG6 and ARG13 have been identified to be the main actors involved in the TAM–SK2 channel interaction.[65−67] Andreotti et al.[65] suggested that these two residues strongly interact with the channel through electrostatics and hydrogen bonding. Furthermore, Ramírez-Cordero et al.[67] showed that mutating one of the three arginines of TAM dramatically decreases its selectivity toward the SK2 channel. It thus appears that the mapping entropy minimization protocol was capable of singling out the two residues that are crucial for a complex biological process. The rationale for this can be found in the fact that such atoms strongly interact with the remainder of the protein, so that small variations of their relative coordinates have a large impact on the value of the overall system’s energy. Retaining these atoms and fixing their positions in the coarse-grained conformation thus enable the model to discriminate effectively one macrostate from another. We note that this result was achieved solely by relying on data obtained in standard MD simulations. This aspect is particularly relevant, as the simulations were performed in absence of the channel, whose size is substantially larger than that of TAM. Consequently, we stress that valuable biological information, otherwise obtained via large-scale, multicomplex simulations, bioinformatic approaches, or experiments, can be retrieved by means of straightforward simulations of the molecule of interest in the absence of its substrate. For AKE (Figure S3), we find that for N = Nα the external, solvent-exposed part of the LID domain is heavily coarse-grained, while its internal region is more conserved. The CORE region of the protein is always largely retained, without noteworthy peaks in probability. Such peaks, on the contrary, appear in correspondence to the terminal nitrogens of ARG36, LYS57, and ARG88 (Pcons(NH2, ARG36) = 0.52; Pcons(NZ, LYS57) = 0.48; Pcons(NH2, ARG88) = 0.58). The two arginines are located in the internal region of the NMP arm, at the interface with the LID domain. ARG88 is known to be the most important residue for catalytic activity,[68,69] being central in the process of phosphoryl transfer.[70] Phenylglyoxal,[71] a drug that mutates ARG88 to a glycine, has been shown to substantially hamper the catalytic capacity of the enzyme.[70] ARG36 is also bound to phosphate atoms.[69] Finally, LYS57 is on the external part of NMP and has been identified to play a pivotal role in collaboration with ARG88 to block the release of adenine from the hydrophobic pocket of the protein.[72] More generally, this amino acid is crucial for stabilizing the closed conformation of the kinase,[73,74] which was never observed throughout the simulation. The overall probability pattern persists as N increases, even though it is less pronounced. For AAT, Figure S3 shows that the associated optimizations heavily coarse-grain the reactive center loop of the protein. On the other hand, two of the most conserved residues in the pool of optimized mappings, MET358 and ARG101, are central to the biological role of this serpin. MET358 (Pcons(CE, MET358) = 0.31) constitutes the reactive site of the protein.[75] Being extremely inhibitor-specific, mutation or oxidation of this amino acid leads to severe diseases. In particular, heavy oxidation of MET358 is one of the main causes of emphysema.[76] The AAT Pittsburgh variant shows MET358ARG mutation, which leads to diminished antielastase activity but markedly increased antithrombin activity.[59,75,77] In turn, ARG101 (Pcons(CZ, ARG101) = Pcons(NH1, ARG101) = Pcons(NH2, ARG101) = 0.35) has a crucial role due to its connection to mutations that lead to severe AAT deficiency.[60,61] In summary, we observe that in all of the proteins investigated, the presented approach identifies biologically relevant residues. Most notably, these residues, which are known to be biologically active in the presence of other compounds, are singled out from substrate-free MD simulations. With the exception of MET358 of AAT, the most probably retained atoms belong to amino acids that are charged and highly solvent-exposed. To quantify the statistical significance of the selection operated by the algorithm, we note that the latter detects those fragments out of pools of 8, 69, and 100 charged residues for TAM, AKE, and AAT, respectively. If we account for solvent exposure, these numbers are reduced to 7, 32, and 40 when amino acids with solvent-accessible surface area (SASA) higher than 1 nm2 are considered. Another aspect worth mentioning is the fact that several atoms pinpointed as highly conserved in optimal mappings are located in the side chains of relatively large residues, such as arginine, lysine, and methionine. It is thus legitimate to wonder whether a correlation might exist between the size of an amino acid and the probability that one or more of its atoms will be present in a low-Smap reduced representation. An inspection of the root-mean-square fluctuation values of the three proteins’ atoms versus their conservation probability (see Figure S4) shows no significant correlation for low or intermediate values of Pcons; highly conserved atoms, on the other hand, tend to be located on highly mobile residues because a relatively large conformational variability is a prerequisite for an atom to be determinant in the mapping. In conclusion, highly mobile residues are not necessarily highly conserved, while the opposite is more likely.

Discussion and Conclusions

In this work, we have addressed the question of identifying the subset of atoms of a macromolecule, specifically a protein, that retains the largest amount of information about its conformational distribution while employing a reduced number of degrees of freedom with respect to the reference. The motivation behind this objective is to provide a synthetic yet informative representation of a complex system that is simulated in high resolution but observed in low resolution, thus rationalizing its properties and behavior in terms of relatively few important variables, namely, the positions of the retained atoms. This goal was pursued by making use of tools and concepts largely borrowed from the field of coarse-grained modeling, in particular bottom-up coarse-graining. The latter term identifies a class of theoretical and computational strategies employed to construct a simplified model of a system that would be too onerous to simulate if it were treated in terms of a high-resolution description. Coarse-graining methods make use of the configurational landscape of the reference high-resolution model to construct a simplified representation that retains its large-scale properties. The interactions among effective sites are parametrized by directly integrating out (in an exact or approximate manner) the higher-resolution degrees of freedom and imposing the equality of the probability distributions of the coarse-grained degrees of freedom in the two representations.[5] These approaches have a long and successful history in the fields of statistical mechanics and condensed matter, the most prominent, pioneering example probably being Kadanoff’s spin block transformations of ferromagnetic systems.[78] This process, which lies at the heart of real-space renormalization group (RG) theory, allows the relevant variables of the system to naturally emerge out of a (potentially infinite) pool of fundamental interactions, thus linking microscopic physics to macroscopic behavior.[79,80] The generality of the concepts of the renormalization group and coarse-graining has naturally taken them outside of their native environment,[81−83] with the whole field of coarse-grained modeling of soft matter being one of the most fruitful offsprings of this cross-fertilization.[5] However, a straightforward application of RG methods in this latter context is severely restricted by fundamental differences between the objects of study. Most notably, the crucial assumptions of self-similarity and scale invariance, which justify the whole process of renormalization at the critical point, clearly do not apply to, say, a protein in that the latter certainly does not resemble itself upon resolution reduction. Furthermore, scaling laws cannot be applied to a system such as a biomolecule that is intrinsically finite, for which the thermodynamic limit is not defined. Additionally, one of the key consequences of self-similarity at the critical point is that the filtering process put forward by the renormalization group turns out to be largely independent of the specific coarse-graining prescription: the set of relevant macroscopic variables emerges as such for almost whatever choice of mapping operator is taken to bridge the system across different length scales.[84] In the case of biological matter, where the organization of degrees of freedom is not fractal but rather hierarchical—from atoms to residues to secondary structure elements and so on—the mapping operator acquires instead a central role in the “renormalization” process. The choice of a particular transformation rule, projecting an atomistic conformation of a molecule to its coarse-grained counterpart, more severely implies an external—i.e., not emergent—selection of which variables are relevant in the description of the system and which others are redundant. In this way, what should be the main outcome of a genuine coarse-graining procedure is demeaned to be one of its ingredients. It is only recently that the central importance of the resolution distribution, i.e., the definition of the CG representation, has gradually percolated into the field of biomolecular modeling.[22,44] Moving away from a priori selection of the effective interaction sites,[21] a few different strategies have been developed that rather aim at the automatic identification of CG mappings. These techniques rely on specific properties of the system under examination: examples include quasi-rigid domain decomposition[24−31] or graph-theory-based model construction methods that attempt to create CG representations of chemical compounds based only on their static graph structure;[32,33,85] other approaches aim at selecting those representations that closely match the high-resolution model’s energetics.[22,35] Finally, more recent strategies rooted in the field of machine learning generate discrete CG variables by means of variational autoencoders.[86] All of these methods take into account the system structure, or its conformational variability, or its energy, but none of them integrates these complementary properties in a consistent framework embracing topology, structure, dynamics, and thermodynamics. In this context, information-theoretical measures, such as the mapping entropy,[17,42−44] can bring novel and potentially very fruitful features.[87] In fact, this quantity associates structural and thermodynamic properties, so that both the conformational variability of the system and its energetics are accurately taken into account. Making use of the advantages offered by the mapping entropy, we have developed a protocol to identify, in an automated, unsupervised manner, the low-resolution representation of a molecular system that maximally preserves the amount of thermodynamic information contained in the corresponding higher-resolution model. The results presented here suggest that the method may be capable of identifying not only thermodynamically consistent but also biologically informative mappings. Indeed, a central result reported here is that those atoms consistently retained with high probability across various lowest-Smap mappings for different numbers of CG sites tend to be located in amino acids that play a relevant role in the function of the three proteins under examination. Most importantly, these key residues, whose biological activity consists of binding with other molecules, have been singled out on the basis of plain MD simulations of the substrate-free molecules in explicit solvent. In general, the vast majority of available techniques for the identification of putative binding or allosteric sites in proteins explicitly or implicitly rely on the analysis of the interaction between the molecule of interest and its partner—be that a small ligand, another protein, or something else.[88−93] This is the case, for example, of binding site prediction servers,[94,95] which perform a structural comparison between the target protein and those archived in a precompiled, annotated database; other bioinformatic tools make use of machine learning methods[96−99]—with all of the pros and cons that come with training over a possibly vast but certainly finite data set of known cases.[100] To the best of our knowledge, the remaining alternative methods perform a structural analysis of the protein in search of binding pockets based on purely geometrical criteria.[101,102] The results obtained in the present work, on the contrary, suggest that a significant fraction of biologically relevant residues, whose function is intrinsically related to interactions with other molecules, might be identified as such from the analysis of simulations in absence of the substrate. This observation would imply that a substantial amount of information about functional residues, even those that exploit their activity through the interaction with a partner molecule, is entailed in the protein’s own structure and energetics. In the past few decades, the successful application of extremely simplified representations of proteins such as elastic network models has shown that the key features of a protein’s large-scale dynamics are encoded in its native structure;[27,36−41,103−107] in analogy with this, we hypothesize that the mapping entropy minimization protocol is capable of bringing to light those relational properties of proteins—namely the interaction with a substrate—from the thermodynamics of the single molecule in the absence of its partner. The mapping entropy minimization protocol establishes a quantitative bridge between a molecule’s representation—and hence its information content—on the one side and the structure–dynamics–function relationship on the other. This method might represent a novel and useful tool in various fields of applications, e.g., for the identification of important regions of proteins, such as druggable sites and allosteric pockets, relying on simple, substrate-free MD simulations and efficient analysis tools. In this study, a first exploration of the method’s capabilities, limitations, and potential developments has been carried out, and several perspectives lie ahead that deserve further exploration. Among the most pressing and interesting ones, we mention the investigation of how the optimized mappings depend on the conformational space sampling; the relation of mapping entropy minimization to more established schemes such as the maximum entropy method; and the viability of a machine-learning-based implementation of the protocol, e.g., making use of deep learning tools that have proven to be strictly related to coarse-graining, dimensionality reduction, and feature extraction. All of these avenues are objects of ongoing study. In conclusion, it is our opinion that the proposed automated selection of coarse-grained sites has great potential for further development, being at the nexus between molecular mechanics, statistical mechanics, information theory, and biology.

Methods

In this section we describe the technical preliminaries and the details of the algorithm we employ to obtain the CG representation M (see eq ) that minimizes the loss of information inherently generated by a coarse-graining procedure—that is, the mapping entropy. Equation provides us with a way of measuring the mapping entropy of a biomolecular system associated with any particular choice of decimation of its atomistic degrees of freedom. One can visualize a decimation mapping (eq ) as an array of bits, where 0 and 1 correspond to not retained and retained atoms, respectively. Order matters: swapping two bits produces a different mapping operator. Applying this procedure, one finds that the total number of possible CG representations of a biomolecule, irrespective of how many atoms N are selected out of n, iswhich is astronomical even for the smallest proteins. In this work, we restrict the set of possibly retained sites to the Nheavy heavy atoms of the compound—excluding hydrogens—thus significantly reducing the cardinality of the space of mappings. Nonetheless, finding the global minimum of eq for a reasonably large molecule would be computationally intractable whenever N is different from 1, 2, Nheavy – 1, and Nheavy – 2. As an example, there are 2.4 × 1038 CG representations of tamapin with 31 sites (N = Nα) and 9.6 × 10887 representations for antitrypsin with 1488 sites (N = Nbkb). Hence, it is necessary to perform the minimization of the mapping entropy through a Monte Carlo-based optimization procedure, and we specifically rely on the simulated annealing protocol.[62,63] As it is typically the case with this method, the computational bottleneck consists of the calculation of the observable (the mapping entropy) at each SA step. We develop an approximate method that is able to obtain the mapping entropy of a biomolecule by analyzing an MD trajectory that can contain up to tens of thousands of frames. At each SA step, that is, for each putative mapping, the algorithm calculates a similarity matrix among all of the generated configurations. The entries of this matrix are given by the root-mean-square deviation (RMSD) between structure pairs, the latter being defined only in terms of the retained sites associated with the CG mapping, and aligned accordingly; we then identify CG macrostates by clustering frames on the basis of the distance matrix, making use of bottom-up hierarchical clustering (UPGMA[108]). Finally, we determine the observable of interest from the variances of the atomistic intramolecular potential energy of the protein corresponding to the frames that map onto the same CG conformational cluster (see eq ). The protocol is initiated with the generation of a mapping such that the overall number of retained sites is equal to N. Then, in each SA step, the following operations are performed:Once the new value of S̃map is obtained, the move is accepted/rejected using a Metropolis-like rule. The overall workflow of the algorithm is illustrated schematically in Figure .
Figure 5

Schematic representation of the algorithmic procedure described in the text that we employ to minimize the mapping entropy, the latter being calculated by means of eq . The full similarity matrix is computed once every TK steps, while in the intermediate steps we resort to the approximation given by eq . TK depends on both the protein and N. TMAX is the number of simulated annealing steps (here TMAX = 2 × 104).

swap a retained site (σ = 1) and a removed site (σ = 0) in the mapping; compute a similarity matrix among CG configurations using the RMSD; apply a clustering algorithm on the RMSD matrix in order to identify the CG macrostates R; compute S̃map using eq . Schematic representation of the algorithmic procedure described in the text that we employ to minimize the mapping entropy, the latter being calculated by means of eq . The full similarity matrix is computed once every TK steps, while in the intermediate steps we resort to the approximation given by eq . TK depends on both the protein and N. TMAX is the number of simulated annealing steps (here TMAX = 2 × 104). For the sake of the accuracy of the optimization, more exhaustive sampling is better, and hence, the number of sampled atomistic configurations should be at least on the order of tens of thousands. However, in that case step 2 would require the alignment of a huge number of structure pairs for each proposed CG mapping, which in turn would dramatically slow down the entire process. This problem is circumvented performing a reasonable approximation in the calculation of the CG RMSD matrix.

RMSD Matrix Calculation

The RMSD between two superimposed structures and is given bywhere n is the number of sites in the system, be they atomistic or CG, and and represent the Cartesian coordinates of the ith elements in the two sets. According to Kabsch,[109,110] it is possible to find the superimposition that minimizes this quantity, namely, the rotation matrix U that has to be applied to for a given in order to reach the minimum of the RMSD. The aforementioned procedure is not computationally heavy per se; in our case, however, we would have to repeat this alignment for all configuration pairs in the MD trajectory every time a new CG mapping is proposed along the Monte Carlo process, thus making the overall workflow inctractable in terms of computational investment. The simplest solution to this problem is to discard the differences in the Kabsch alignment between two CG structures differing by a pair of swapped atoms. This assumption is particularly appealing from the point of view of speed and memory, since the expensive and relatively slow alignment procedure produces a result (a rotation matrix) that can be stored with negligible use of resources. In order to take advantage of this simplification without losing accuracy, for each structure and degree of coarse-graining we select an interval of SA steps TK in which we consider the rotation matrices constant. After these steps, the full Kabsch alignment is applied again. This approximation results in a substantial reduction in the number of operations that we have to execute at each Monte Carlo step. At first, given the initial random mapping operator M, we build the sets of coordinates that have been conserved by the mapping operator Γ(M) = M(r). Then we compute RMSD(Γα(M), Γβ(M)), the overall RMSD matrix between every pair of aligned structures Γα and Γβ, where α and β run over the MD configurations. For all moves M → M′ within a block of TK Monte Carlo steps, where M and M′ differing only in a pair of swapped atoms, this quantity is then updated with the simple rulewhere and are the removed (substituted) and added atoms, respectively, and MSD is the mean-square deviation. This approach clearly represents an approximation to the correct procedure; it has to be emphasized, however, that the impact of this approximation is increasingly perturbative as the size of the system grows. Furthermore, the computational gain that the described procedure enables is sufficient to counterbalance the fact that the exact protocol would be so inefficient to make the optimization impossible. For example, with TK = 1000 for AAT with N = Nbkb, our approximation gives a speed-up factor of the order of 103.

Hierarchical Clustering of Coarse-Grained Configurations

Several clustering algorithms exist that have been applied to group molecular structures based on RMSD similarity matrices.[111,112] Many such algorithms have been developed and incorporated in the most common libraries for data science. Among the various available methods, we choose to employ the agglomerative bottom-up hierarchical clustering with average linkage (UPGMA) algorithm.[108] Here we briefly recapitulate the basics underpinnings of this procedure: In the first step, the minimum of the similarity matrix is found, and the two corresponding entries x, y (leaves) are merged together in a new cluster k. k is placed in the middle of its two constituents. The distance matrix is updated to take into account the presence of the new cluster in place of the two close structures: d(k, z) = (d(x, z) + d(y, z))/2. Steps 1 and 2 are iterated until one root is found. The distance among clusters k and w is generalized as follows:where |k| and |w| are the populations of the clusters and k[i] and w[j] are their elements. The actual division into clusters can be performed by cutting the tree (dendrogram) using a threshold value on the intercluster distance or taking the first value of the distance that gives rise to a certain number of clusters Ncl. In both cases it is necessary to introduce a hyperparameter. In our case, the latter is a more viable choice to reduce the impact of roundoff errors. Indeed, the first criterion would push the optimization to create as many clusters as possible in order to minimize the energy variance inside them (a cluster with one sample has zero variance in energy). This algorithm, whose implementation[113,114] is available in Python Scipy,[115] is simple, relatively fast (O(n2 log n)), and completely deterministic: given the distance matrix, the output dendrogram is unique. Although this algorithm scales well with the size of the data set, it may not be robust with respect to small variations along the optimization trajectory. In fact, even the slightest modifications of the dendrogram may lead to abrupt changes in S̃map. This is perfectly understandable from an algorithmic point of view, but it is deleterious for the stability of the optimization procedure. Furthermore, the aforementioned choice of Ncl is somehow arbitrary. Hence, we perform the following analysis in order to enhance the robustness of S̃map at each Monte Carlo (MC) move and to provide a quantitative criterion to set the hyperparameter: compute the RMSD similarity matrix between all of the heavy atoms of the biological system under consideration; apply the UPGMA algorithm to this object, retrieving the all-atom dendrogram; impose lower and upper bounds on the intercluster distance depending on the conformational variability of the structure (see Table );
Table 3

Bounds on Intercluster Distances and Corresponding Numbers of Clusters

proteinupper bound (nm)lower bound (nm)Ncl+Ncl
TAM0.200.189134
AKE0.250.2014729
AAT0.200.15967
visualize the cut dendrogram to identify the numbers of different clusters available at each of the two threshold values (Ncl+ and Ncl–; Table ); build a list CL of five integers, selecting three (intermediate) values between Ncl– and Ncl+; define the observable as the average over the values of S̃map (see eq ) computed choosing different Ncl values:where |CL| is the cardinality of the chosen list. The overall procedure amounts to identifying many different sets of CG macrostates R on which S̃map can be computed, assuming that the average of this quantity can be used effectively as the driving observable inside the optimization. This trivial assumption increases the robustness of the SA optimization and allows all of the values of S̃map calculated at different distances from the root of the dendrogram to be kept in memory.

Simulated Annealing

We use Monte Carlo simulated annealing to stochastically explore the space of the possible decimation mappings associated with each degree of coarse-graining. We here briefly describe the main features of our implementation of the SA algorithm, referring the reader to a few excellent reviews for a comprehensive description of the techniques that can be employed in the choice of temperature decay and parameter estimation.[116,117] We run the optimization for 2 × 103 MC epochs, each of which is composed of 10 steps. This amounts to keeping the temperature constant for 10 steps and then decreasing it according to an exponential law. For the ith epoch, we have T(i) = T0e–. The hyperparameters T0 and ν are crucial for a well-behaved MC optimization. We choose ν = 300 so that the temperature at i = 2000 is approximately T0/1000. In order to feed our algorithm with reasonable values of T0, for each of 100 random mappings we perform 10 MC stochastic moves and measure ΔΣ, the difference between the observables computed in two consecutive steps. Then we estimate T0 so that a move that leads to an increment of the observable equal to the average of ΔΣ would possess an acceptance probability of 0.75 at the first step.

Data Available

For each analyzed protein, the raw data for all of the CG representations investigated in this work, including random, optimized, and transition mappings, together with the associated mapping entropies are freely available on the Zenodo repository (https://zenodo.org/record/3776293). We further provide all of the scripts we employed to analyze such data and construct all of the figures presented in this work.
  89 in total

1.  Anisotropy of fluctuation dynamics of proteins with an elastic network model.

Authors:  A R Atilgan; S R Durell; R L Jernigan; M C Demirel; O Keskin; I Bahar
Journal:  Biophys J       Date:  2001-01       Impact factor: 4.033

2.  FTSite: high accuracy detection of ligand binding sites on unbound protein structures.

Authors:  Chi-Ho Ngan; David R Hall; Brandon Zerbe; Laurie E Grove; Dima Kozakov; Sandor Vajda
Journal:  Bioinformatics       Date:  2011-11-22       Impact factor: 6.937

3.  PiSQRD: a web server for decomposing proteins into quasi-rigid dynamical domains.

Authors:  T Aleksiev; R Potestio; F Pontiggia; S Cozzini; C Micheletti
Journal:  Bioinformatics       Date:  2009-08-20       Impact factor: 6.937

4.  The Renormalization Group and Its Applications to Generating Coarse-Grained Models of Large Biological Molecular Systems.

Authors:  Patrice Koehl; Frédéric Poitevin; Rafael Navaza; Marc Delarue
Journal:  J Chem Theory Comput       Date:  2017-02-24       Impact factor: 6.006

5.  Dual-potential approach for coarse-grained implicit solvent models with accurate, internally consistent energetics and predictive transferability.

Authors:  Kathryn M Lebold; W G Noid
Journal:  J Chem Phys       Date:  2019-10-28       Impact factor: 3.488

6.  Molecular modeling and docking simulations of scorpion toxins and related analogs on human SKCa2 and SKCa3 channels.

Authors:  Nicolas Andreotti; Eric di Luccio; François Sampieri; Michel De Waard; Jean-Marc Sabatier
Journal:  Peptides       Date:  2005-04-19       Impact factor: 3.750

Review 7.  Scorpion venom components that affect ion-channels function.

Authors:  V Quintero-Hernández; J M Jiménez-Vargas; G B Gurrola; H H Valdivia; L D Possani
Journal:  Toxicon       Date:  2013-07-26       Impact factor: 3.033

8.  The comparison of automated clustering algorithms for resampling representative conformer ensembles with RMSD matrix.

Authors:  Hyoungrae Kim; Cheongyun Jang; Dharmendra K Yadav; Mi-Hyun Kim
Journal:  J Cheminform       Date:  2017-03-23       Impact factor: 5.514

9.  BioLiP: a semi-manually curated database for biologically relevant ligand-protein interactions.

Authors:  Jianyi Yang; Ambrish Roy; Yang Zhang
Journal:  Nucleic Acids Res       Date:  2012-10-18       Impact factor: 16.971

10.  Mechanical and assembly units of viral capsids identified via quasi-rigid domain decomposition.

Authors:  Guido Polles; Giuliana Indelicato; Raffaello Potestio; Paolo Cermelli; Reidun Twarock; Cristian Micheletti
Journal:  PLoS Comput Biol       Date:  2013-11-14       Impact factor: 4.475

View more
  6 in total

Review 1.  Bottom-up Coarse-Graining: Principles and Perspectives.

Authors:  Jaehyeok Jin; Alexander J Pak; Aleksander E P Durumeric; Timothy D Loose; Gregory A Voth
Journal:  J Chem Theory Comput       Date:  2022-09-07       Impact factor: 6.578

2.  A new one-site coarse-grained model for water: Bottom-up many-body projected water (BUMPer). II. Temperature transferability and structural properties at low temperature.

Authors:  Jaehyeok Jin; Alexander J Pak; Yining Han; Gregory A Voth
Journal:  J Chem Phys       Date:  2021-01-28       Impact factor: 3.488

Review 3.  Bottom-Up Coarse-Grained Modeling of DNA.

Authors:  Tiedong Sun; Vishal Minhas; Nikolay Korolev; Alexander Mirzoev; Alexander P Lyubartsev; Lars Nordenskiöld
Journal:  Front Mol Biosci       Date:  2021-03-17

Review 4.  Fifty years of Shannon information theory in assessing the accuracy and agreement of diagnostic tests.

Authors:  Alberto Casagrande; Francesco Fabris; Rossano Girometti
Journal:  Med Biol Eng Comput       Date:  2022-02-23       Impact factor: 2.602

Review 5.  "Dividing and Conquering" and "Caching" in Molecular Modeling.

Authors:  Xiaoyong Cao; Pu Tian
Journal:  Int J Mol Sci       Date:  2021-05-10       Impact factor: 5.923

6.  A journey through mapping space: characterising the statistical and metric properties of reduced representations of macromolecules.

Authors:  Roberto Menichetti; Marco Giulini; Raffaello Potestio
Journal:  Eur Phys J B       Date:  2021-10-12       Impact factor: 1.500

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.