- Methodology article
- Open Access
A maximum likelihood framework for protein design
BMC Bioinformatics volume 7, Article number: 326 (2006)
The aim of protein design is to predict amino-acid sequences compatible with a given target structure. Traditionally envisioned as a purely thermodynamic question, this problem can also be understood in a wider context, where additional constraints are captured by learning the sequence patterns displayed by natural proteins of known conformation. In this latter perspective, however, we still need a theoretical formalization of the question, leading to general and efficient learning methods, and allowing for the selection of fast and accurate objective functions quantifying sequence/structure compatibility.
We propose a formulation of the protein design problem in terms of model-based statistical inference. Our framework uses the maximum likelihood principle to optimize the unknown parameters of a statistical potential, which we call an inverse potential to contrast with classical potentials used for structure prediction. We propose an implementation based on Markov chain Monte Carlo, in which the likelihood is maximized by gradient descent and is numerically estimated by thermodynamic integration. The fit of the models is evaluated by cross-validation. We apply this to a simple pairwise contact potential, supplemented with a solvent-accessibility term, and show that the resulting models have a better predictive power than currently available pairwise potentials. Furthermore, the model comparison method presented here allows one to measure the relative contribution of each component of the potential, and to choose the optimal number of accessibility classes, which turns out to be much higher than classically considered.
Altogether, this reformulation makes it possible to test a wide diversity of models, using different forms of potentials, or accounting for other factors than just the constraint of thermodynamic stability. Ultimately, such model-based statistical analyses may help to understand the forces shaping protein sequences, and driving their evolution.
Predicting the sequences compatible with a given structure defines what is traditionally called the inverse folding problem, or more often, protein design [1–3]. As suggested by the terminology, this question is usually considered in an engineering perspective: the aim is then to determine a sequence, or a set of sequences, that stably fold into a pre-specified conformation. In a thermodynamic perspective, this requirement translates into eliciting sequences that have lowest free energy under the target fold, compared to all possible alternative conformations. In principle, such a criterion would imply a search through the joint structure-sequence space, which is not feasible but for small on-lattice model proteins .
As an alternative to the engineering approach, a more evolutionary stance can be taken towards the inverse folding problem, in which case the aim would rather be to predict the sequences of natural proteins having the conformation of interest. Seen from this new point of view, the design problem raises new questions: natural proteins are the result of a complex evolutionary process, involving an intricate interplay between mutation and selection, and this probably entails many constraints directly related to the native conformation, but nevertheless not equivalent to the mere requirement of structural stability. For instance, the requirement of fast and cooperative folding has an impact on the dispersion of contact energies . For this and many other potential reasons, among all sequences predicted by classical engineering-oriented protein design, probably only a subset will look like natural proteins.
The evolutionary approach to protein design is particularly relevant to phylogenetic studies, where one of the current motivations is to develop the so-called structurally constrained models of protein evolution, i.e. models explicitly dependent on the protein's conformation, either for simulation purposes [6–9], or in the context of phylogenetic inference [10, 11]. In this framework, each substitution undergone by a protein during evolution has to be tested for its compatibility with the structure, in the context of the sequence that the protein displays at all other sites when the substitution occurs. Such repeated evaluation of the structure-sequence compatibility along a phylogenetic tree requires relevant and computationally very efficient scoring schemes/functions.
It is interesting to compare the different methods proposed thus far for performing protein design in light of this engineering/evolutionary distinction. A first direction of research has consisted in using all-atom semi-empirical force fields to evaluate the conformational free energy (reviewed in ). These empirical methods have been applied to many theoretical and experimental cases, reaching a high level of accuracy. On the other hand, they are computationally heavy, mainly because of the side-chain positioning problem, and thus cannot be easily applied to structurally constrained phylogenetic models [10, 11]. Concerns may also be expressed about their over-sensitivity to the native conformation, in particular in the core of the target structures and when the flexibility of the backbone is not accounted for [13, 14]. But more importantly, approaches based on physical force fields are, by definition, exclusively focussed on the conformational stability, and thereby, completely oversee other potential factors shaping the sequences of biological proteins. As such, they are well suited for engineering synthetic proteins , or for testing to what extent natural sequences are shaped by selection for protein stability , but may not be sufficient for more general evolutionary purposes.
An alternative to the semi-empirical strategy consists in relying on knowledge-based, or statistical, potentials. These scoring functions mimic physical Boltzmann distributions, but merely encode statistical patterns present in the databases. Some of these potentials were obtained under the quasi-chemical approximation, whereby frequencies of patterns, such as contacts between each pair of amino-acids, are transformed into energies using the Boltzmann law [17–20]. Alternatively, contact energies can be obtained by maximizing the potential's predictive accuracy in a threading test [21–24]. In the present context, an advantage of these knowledge-based potentials, compared to semi-empirical force-fields, is that they should in principle capture all kinds of patterns that true biological sequences have, in relation to their conformation, and not only those directly related to thermodynamic stability. Furthermore, statistical potentials need not be defined at the atomic level, but can be based on a coarse-grained description of the protein's configuration, essentially by omitting the degrees of freedom associated to side chains. This allows faster computations, by avoiding the problem of searching through the rugged landscape of side-chain conformations. In addition, coarse-grained potentials could turn out to be an advantage, in that they will not recover the native sequence too faithfully. Most protein design procedures based on statistical potentials proposed until now have relied on coarse-grained, pairwise contact pseudo-energies [4, 25–32].
Yet, irrespective of the level of description adopted, currently available statistical potentials may not be ideal for protein design, since they have generally been optimized in the context of the folding problem, i.e. for maximizing the rate of correct structure prediction, given the sequence. In contrast, we would like to optimize the reciprocal prediction, namely, the sequences given the conformation. Several approaches have been proposed in this direction, consisting in maximizing the Z-score between the energy of the native sequence on the target conformation and its energy on a set of decoy sequences , or, alternatively, in applying a mean-square criterion on the values taken by the scoring function on each structure-sequence pair of the database . However, these methods have thus far only been tested in cubic lattice protein models. In addition, they lack a firm theoretical basis. In particular, it would be interesting to guarantee optimal predictive power, and to have a robust methodology available to assess and compare the performance of alternative forms of statistical potentials.
Standard statistical theory provides such theoretical guarantees . In the present case, the inverse folding problem can be formulated directly in terms of the probability of observing a sequence s given a conformation c, i.e. p(s | c, θ). This probability explicitly depends on the pre-specified model through a series of parameters, represented here by θ. These may be, for instance, the coefficients of a pairwise potential, parameters describing compositional effects, secondary structure environment, solvent accessibility, etc. Taking the product over a database of P independent sequence-conformation pairs, S = (sp)p = 1..Pand C = (cp)p = 1..p, yields a joint probability
which, as a function of θ, can be seen as a likelihood. The parameter θ is then learnt by maximizing the likelihood with respect to θ. Once this is done, sequences can be assessed, or sampled, under the optimal parameter value , by direct numerical evaluation of their probability, or by Monte Carlo sampling methods.
Reformulated in this way, the method maximizes the predictive power of the potential, now in the structure-seeks-sequence direction. By construction, it yields the optimal parameter values that can be obtained for a given form of the potential. In addition, the fit of the model can be directly evaluated, based on the value of the likelihood obtained on a test data set, distinct from the learning set (cross-validation), giving a means of rigorous model selection. Finally, the statistical framework proposed here allows one to explicitly combine together, in a model dependent manner, all kinds of factors that we surmise may induce correlations between the structure and the sequence of proteins.
We have implemented this maximum likelihood (ML) procedure in a Markov chain Monte Carlo framework, and applied it to a simple case, using a contact potential, supplemented with a solvent accessibility term. Using cross-validation, we show that the resulting potentials yield a better fit than currently available potentials of the same form, and that combining solvent-accessibility considerations with contact energies is better than either alone. Furthermore, we find that solvent accessibility requires a more complex description than what is currently used. Ultimately, the overall method proposed in this work can be extended to a large spectrum of alternative models and statistical potentials.
The probabilistic model
Let us consider a sequence s = (s i )i = 1..N, of length N, and of conformation c. In its most general form, the method introduced here can work with any model M specifying the conditional probability of s given c, in terms of an unnormalized non negative function q(s, c):
To illustrate the method, we will apply it to a simple case, using a pairwise contact potential. The argument is as follows. First, by Bayes' theorem:
If, in addition, we assume a uniform prior on s, we can simply relate equations 3 and 2 by posing q(s, c) = p(c | s, M). Next, given a statistical potential E(s, c), the conformational probability p(c | s) can be expressed as a Boltzmann distribution:
is a normalization constant, and
F(s) = - ln Z s . (7)
T and k are the absolute temperature and the Boltzmann constant, respectively. Without loss of generality, it is possible to rescale the potential so that kT = 1, which we will do in the following.
Then, by defining the inverse potential:
G(s, c) = E(s, c) - F(s), (8)
the conditional probability of sequence s reads as
is the normalization factor. Note that, contrary to the Z s factor of equation 4, which was a sum over all conformations, the present factor Y is a sum over sequence space (all possible sequences of length N).
In the present work, we used a statistical potential made of two terms:
The first term is a contact free energy: Δ ij = 1 if positions i and j are closer in space than a certain cut-off distance, and 0 otherwise, and ε ab defines the contact energy between amino acids a and b. The second term encodes a solvent-accessibility free energy: for each position, represents the free energy of amino acid a in the solvent accessibility class d, a = 1..20, and d = 1..D, where D is the total number of solvent accessibility classes considered.
Deriving the inverse potential requires the calculation of F(s), which is already entirely specified by the potential E as a sum over all conformations. However, this computation is difficult in practice. As an alternative, we can give it a simple phenomenological form, inspired from the random energy model [25, 28, 35]:
where the (μ a )a = 1..20are unknown parameters, analogous to "chemical potentials" for the 20 amino acids.
Altogether, our parameter vector is made of three components: θ = (α, ε, μ), and the inverse potential reads as:
Note that the probability defined by equation 9 is invariant under the following transformation:
= μ a + J1, (14)
= ε ab + J2, (15)
= + J3, (16)
where J1, J2 and J3 are arbitrary real constants. Therefore, to ensure identifiability of our probabilistic model, we enforce the following constraints:
A series of alternative inverse potentials can be obtained by suppressing the first or the second of the components of equation 13. In the present work, we tested the following combinations:
α + μ,
ε + μ,
ε + α + μ.
We also explored various numbers of accessibility classes, with D ranging from 2 to 20. Alternatively, the ε component can be fixed to values of a contact potential obtained by other authors (MJ) . In this case, we must add a multiplicative scaling factor λ in front of the contact component to account for the fact that these potentials are normalized differently:
The scaling factor is optimized by ML, along with μ.
Optimizing the potentials by gradient descent
If we now consider a database, made of P protein sequences S = (sp)p = 1..P, of respective lengths N p and their corresponding three dimensional structures C = (cp)p = 1..P, the probability of observing the whole database, which we define as the likelihood L(θ), is the product of the probabilities of observing each protein independently:
is the inverse potential summed over the database, and
is the corresponding normalization constant. Since it is more convenient to work on minus the logarithm of the probability, we define the score ω:
ω(θ) = - ln L(θ) (26)
= G(S, C) + ln Y. (27)
We wish to maximize the likelihood, or equivalently, minimize ω, with respect to θ. We do this by gradient descent, based on a numerical evaluation of the derivative of ω (see methods). The overall method is akin to an Expectation Maximization algorithm . In fact, it can be seen as a differential version of Dempster's method, and therefore, we call it differential EM.
The derivative of ω reads as:
Applying the partition function formalism to equation 25, we can express the second term as an expectation over p(S' | C, θ):
which leads us to the following expression for the derivative of ω:
The computation of the first term in this equation is straightforward, while the second term must be estimated numerically. In order to do so, we obtain a sample
drawn from p(S | C, θ) by a Gibbs sampling algorithm similar to that of Robinson et al.  (see methods).
Applying formula 33 on the inverse potential 13 yields the following expressions for the derivatives:
where n ab is the number of contacts between amino acids a and b observed in the database, and 〈n ab 〉 is its expectation over the probability distribution p(S' | C, θ). Formula 34 thus leads to an intuitive characterization of the maximum likelihood estimate : it is the value of ε such that the average number of each type of contact predicted by the potential matches the number observed in the database. Following a similar derivation:
where m a is the total number of amino acids of type a, and
where is the total number of amino acids of type a belonging to solvent-accessibility class d.
We first performed an optimization of the pure contact potential (ε + μ-potential) on each data set. Figure 1 shows the evolution of the scoring function ω and of the contact potential during the gradient descent. As can be seen from these traceplots, the differential EM algorithm converges after a few hundred cycles. The scoring function stabilizes at around 272,000 natural units of logarithm (nits), and then fluctuates by up to 25 nits around this value. These fluctuations are mainly due to the finite size of the sample of sequences on which the derivative of ln Y is evaluated and, to a lesser extent, to the error on the estimation of ln Y by thermodynamic integration. In any case, these errors are small compared to the differences between scores obtained with alternative models (see below).
The evolution of the potential for some residue pairs is shown in figure 1b and 1c. Effects in the final values due to residue polarity are easily seen: known favorable interactions such as glutamate-lysine or the hydrophobic isoleucine-valine have a lower contact energy, while known unfavorable interactions, such as glutamate-glutamate, have higher energies, indicating that the potentials obtained are biologically reasonable.
The potentials obtained in two independent runs are virtually identical (figure 2a), indicating that the gradient descent does not get trapped into local minima. We can also compare the values of the potential for two distinct data sets of equivalent size, DS1 and DS2 (figure 2b), which uncovers a greater discrepancy than for two independent runs on the same data set DS1. The correlation is high, however, suggesting that data sets are large enough for the learning procedure to reach stability. In addition, these differences are small compared to the discrepancy between the potential obtained by our method and that of Miyazawa & Jernigan (figure 2c).
The same optimization procedure was applied to the potential consisting only of the solvent accessibility term (α + μ), with an increasing number of accessibility classes, and to the combined (ε + α + μ) potential. The resulting log likelihood scores cannot directly be compared, since the models do not have the same dimensionality. We therefore applied a 2-fold cross-validation procedure (CV), consisting in learning the potential on DS2, and testing it on DS1, and vice versa.
The evolution of the CV score as a function of the number of accessibility classes (D) is shown in figure 3. When D increases, the fit of the model improves, until a point is reached where the penalization for model dimensionality starts to dominate the score. The optimal number of classes obtained is 14 to 16, depending on the form of the potential studied, although 4 to 6 classes is sufficient to attain 90% of the fit improvement.
The scores obtained for the different models tested are reported in figure 4. We also included in the comparison the Miyazawa and Jernigan potential . The contact potential performs better than the pure solvent accessibility potential, and the combination of both terms is the most informative. Miyazawa and Jernigan's potential results in a poorer fit improvement than any of the other models.
Specificity of the designed sequences
Once an optimal value of θ is obtained, properties of the sequences induced by the models can be investigated by sampling sequences from p(s | c, θ), using this optimal value of θ. In particular, we tested to what extent the sequences proposed by our method met the requirement of specificity, i.e. the condition that the sequences designed on a given conformation c indeed have c as their unique ground state. More precisely, we generated 20 sequences by Gibbs sampling for 60 randomly chosen structures [see Additional file 8], i.e. 1,200 sequences for each potential, and performed a fold recognition experiment for the designed sequences, monitoring the score for the target fold using THREADER  (figure 6 and Table 1).
The solvent accessibility potential alone (α14ac + μ, figure 6b) is not sufficient to provide specificity to the designed sequences, and behaves almost as poorly as the flat potential (μ, figure 6a). A mild improvement is seen when using the contact potential (ε + μ, figure 6c): for 10% of the designed sequences the target fold is found among the best scoring folds (Table 1), and the distribution of this ranking is skewed towards lower values. However, it is only with the combined potential (ε + μ14ac + μ, figure 6d) that a significant improvement is observed: for more than half of the designed sequences the target fold is found among the best 1% scoring folds, even though the average sequence identity with the native sequence is less than 10% in all cases (Table 1).
We also tested a subset of 120 randomly chosen designed sequences using another fold recognition program, LOOPP . LOOPP is based on a combination of several structure prediction methods, based on threading, secondary structure, sequence profile and exposed surface area prediction. The results obtained with this program were similar to those of THREADER: for 51.2% of the designed sequences using the combined (ε + α14ac + μ) potential, the target fold was found as the first hit, and for 67.2% the target fold was found among the first 10 hits.
In contrast, many of the current fold recognition programs based on sequence profile methods produced no significant hits (data not shown), which is not surprising, given that our sampling algorithm produces highly divergent sequences, with no similarity to any natural protein.
The central idea of the present work is to reformulate the problem of devising statistical potentials for protein design as a statistical inference problem. This reformulation, based on the maximum likelihood (ML) principle, led us naturally to a gradient descent method, with the only additional aspect being that the gradient to follow is itself estimated by Monte-Carlo averaging.
The main advantage of this ML framework is that it guarantees an optimal predictive power of the resulting potential. In addition, it is very general, and can in principle be applied to any form of statistical potential. In particular, it is not restricted to coarse grained descriptions of proteins, and it could also be applied at the atomic level.
Interestingly, our gradient descent method turns out to be similar in spirit to an iterative scheme proposed by Thomas and Dill , although in that case the purpose was to optimize a potential in the context of the folding problem. Specifically, Thomas and Dill tune the potential so as to match the observed and expected number of contacts of each type, except that their expectation is taken on a set of alternative conformations, for a fixed sequence, whereas we take the expectation on a set of alternative sequences, on the conformation of interest. Note that Thomas and Dill derived their method from intuitive arguments, and not as a mathematical consequence of the ML principle.
These two alternative optimization schemes, obtained by normalizing either over the sequence or over the structure space, are quite distinct, at least conceptually. How the resulting potentials would differ in practice is more difficult to evaluate. Among other things, it will depend on how the approximation of lnZ s based on the random energy model works. In the eventuality that it does not work well, it is likely that the contact term of our inverse potential will in fact combine two things: the information corresponding to the conformational energy of the sequence itself, which is also encoded in classical potentials optimized for threading, plus some information coming from the decoy term ln Z s . A way to settle this question would be to optimize a contact potential using, on the same learning set, both normalization schemes, and then compare the resulting values as well as their predictive powers.
Model assessment and comparison
The methodological framework proposed here offers reliable criteria for comparing the empirical fit of alternative models on real data. In this respect, it should be noted that the lack of a reliable objective criterion for evaluating different statistical potentials has often been invoked for justifying the use of on-lattice idealized models . However, on-lattice approaches are only moderately interesting, as they completely ignore the problem of the robustness of the learning method to model violation. Coarse-grained statistical potentials are by definition over-simplified models of proteins, and therefore, model violation is an intrinsic feature of the protein design problem. In this respect, the statistical language is interesting, since it is still valid, even for fitting and assessing models that are known to be imperfect.
On the other hand, the intuitive idea underlying cross-validation, i.e. measuring the rate of prediction of the native sequence, is quite simple, and has been invoked and used several times previously [16, 29, 32, 35, 40]. What we propose here is a better formalization of this idea. Note that in contrast to previous methods, we do not measure the marginal native prediction rate at each site, but the joint probability of the native sequence. This can be important, as it accounts for possible correlations in the predictive distribution. For instance, two given positions may not display any particular pattern, when considered marginally, but may jointly follow charge or steric compensatory patterns. These phenomena will not be taken into account in the overall fit of the potential when measuring the marginal prediction rate, as is usually done. Technically speaking, the joint probability of the native sequence on the corresponding structure is extremely small, and cannot be evaluated just by counting the frequency at which the native sequence appears in the sample obtained by Gibbs sampling. For this, more elaborate numerical methods, such as thermodynamic integration, are required.
In the present case, the comparison between alternative models has allowed us to measure the relative contribution of each term of the potential and to refine the protein representation. The contact component turns out to be the most informative (figure 4), although it should be complemented with other energetic forms. Here, we have tested the addition of a solvent accessibility component, which significantly improves the fit of the model. Contact information and solvent exposure are correlated, which is reflected in the fact that the fit improvement of each term is not additive.
Our model comparison method also gives us a direct way of choosing the optimal number of solvent accessibility classes (figure 3). Here, we found a number of 14 to 16 classes, which is higher than what one may have expected and than what is usually used. Note that this number depends on the way the classes are defined; here, the classes are based on quantiles, but as an alternative, we also tried a linear definition (evenly splitting the whole range of accessibility surfaces into D bins), which gave us an even higher optimal number of classes (20 classes, data not shown). In general, the present methodology could be used to investigate different definitions of accessibility classes, to refine the pairwise contact definition, or any other elements of the structure representation included in the potential.
The fact that our potential has a significantly better predictive power than that of Miyazawa and Jernigan (MJ, figure 4) is trivially expected, by construction of the ML potential. What is more surprising is that the MJ matrix is less fit than a simple solvent-accessibility profile. A possible explanation would be that Miyazawa and Jernigan's potential is based on the quasi-chemical approximation, which is now known to be somewhat drastic [19, 41, 42], as it neglects correlations between observed pairing frequencies, due to chain connectivity and multiple contacts. Alternatively, it could mean that potentials optimized for folding are really not suited for protein design purposes. Testing other pairwise contact potentials, in particular those that do not rely on the quasi-chemical approximation [22, 24, 43–45], would be a way to address this issue.
The method that we propose in this work is probabilistic in essence. As such, it offers a very natural framework for investigating the patterns induced by the models on distributions of sequences.
Specificity of the designed sequences
A sequence s designed for a target conformation c should not only be compatible with c, but also incompatible with competing folds. A rigorous solution to this problem involves a simultaneous search over the sequence and conformation space. It is possible, however, to achieve specificity without explicitly seeking to penalize competing states (negative design), if we rely on the approximation based on the random energy model, where the normalization constant of equation 4 can be considered as a function of the sequence composition only [25, 46]. In our case, the normalization of the likelihood will also play an important role: since the total probability over all possible sequences has to be 1, maximizing the probability for a given sequence s1 on its native conformation c1 will lower the probability that another natural sequence s2, with native conformation c2, also gets a high probability on c1. When many sequences are learnt in parallel, this phenomenon should ultimately favor specificity of s2 on c2, compared to all other conformations of the data set.
On the other hand, the extent to which the specificity is achieved will depend on the actual form of the potential used, as well as on the data base used for learning. To address this question, we produced a large number of sequences with four different potentials, and checked their ability to recognize the target fold, as measured by the Z-score ratio or by the ranking of the target structure in a fold recognition experiment. Indeed, an improvement of specificity is observed when using better potentials, suggesting that the method is effectively capturing specific dependencies between the conformation and the sequence of the proteins in the learning set, even for the simple forms of potentials tested here. For the combined (ε + α14ac + μ) potential, the average Z-score ratio of the designed sequences is similar to what has been reported for other protein design algorithms . Conversely, this also suggests that a more sophisticated potential may further improve the specificity of the sequences designed using our algorithm.
Conformation-dependent site-specific profiles
To compare natural protein sequences with those predicted by the optimized potentials, marginal, leave-one-out and empirical profiles (see methods) were generated for the 60 proteins used in the design specificity experiment described above; the profiles obtained for the best and the worst scoring structures are provided as supplementary materials [see Additional file 7]. Overall, leave-one-out profiles (figure 5a) and marginal profiles (figure 5b) do not display significant differences in the discriminative power between sites: the mean Shannon entropy per site is 0.743 ± 0.366 for marginal profiles, and 0.696 ± 0.428 for leave-one-out profiles. It is worth noting that the mean entropy per site for each protein, and the corresponding standard deviation, i.e. the average amount of information at each site and the variation between sites, are both correlated with the performance of the particular protein in the fold recognition experiment, and this, only for the combined (ε + α14ac + μ) potential (Table 1).
A detailed analysis of the leave-one-out profiles for a particular case, an alpha-aminotransferase, may be useful to understand which type of information is effectively captured by the potential, and which is not captured at all, thereby suggesting possible ways of improving the current form of potential.
First, regions of the protein that show little secondary structure (such as in positions 32–40, 55–65 and 82–88) contain less information (mean entropy per site = 0.756) than regions with local structure (mean entropy per site = 0.856). This is not surprising, since these regions typically have fewer contacts between residues, and thus the amount of information included in the protein representation is lower.
Concerning regions with defined secondary structure, residue polarity is the information most easily captured. Charged residues are also distinctively inferred, as well as glycines, to a lesser extent (e.g. glycine 64, 81 and 95 – the latter predicted at position 94 or 95). In contrast, prolines are rarely correctly predicted, which is expected, since the properties most distinctive of prolines (such as phi-psi dihedral angles or local secondary structure) are not included in this particular form of potential.
Interestingly, some residues that have a crucial importance for the protein structure or function fail to be predicted, simply because the properties conferring their importance are not included in the protein description. This is the case of the amino acids that are in close interaction with a ligand (positions 34, 59, 96, 97).
Finally, the leave-one-out profiles display an interesting behavior with respect to positions where the amino-acid present in the reference sequence is not at all conserved in other members of the family. In some cases, they simply do not predict anything (e.g. glycines 24 and 60, or leucine 9, isoleucine 21, and alanine 23), which suggests that their limited importance in structure stability or function is recognized by the inverse potential. In other cases, the natural profile is even reproduced in the leave-one-out profile, instead of the amino acid of the reference sequence; such is the case for phenylalanine 100.
As illustrated by the sequence logos and the fold recognition experiments performed above, the predictive power of the models proposed here is encouraging, but nevertheless still weak. It is not yet clear to what extent this is due to the specific choice made concerning the form of the statistical potential, to the approximation of ln Z s as a function of the sole composition of the sequence, or to yet other reasons. Most probably, we are facing a combination of several factors. The methods proposed here can now be used to address these difficult questions empirically.
In one direction, other approximations of ln Z s , less drastic than the random energy model, but still accessible in practice, can be investigated. For instance, following Deutsch and Kurozky (1996), the conditional probability of a sequence could be defined as:
p(s | c) ∝ e-[E(S,C)-〈E(S)〉]p(s) (37)
where the expectation 〈·〉 is taken over a pre-defined set of decoy conformations. More sophisticated Monte Carlo methods, jointly sampling the sequence and conformation spaces, can also be imagined, in order to get more precise evaluations of ln Z s , while staying in the same global maximum likelihood formalism.
On the other hand, all the many statistical potentials that have been proposed over the last fifteen years may in principle be investigated in the same way as we have done here. In particular, distance-dependent potentials  and main-chain dihedral angle potentials , which imply a richer representation of the protein structure, may result in models of greater predictive power. Other ways of implicitly considering side-chain conformation may also be easily incorporated into the model.
In a completely different perspective, it is possible to devise probabilistic models that are not exclusively defined in terms of a conformational free energy, even in a formal way. For instance, additional terms, concerning secondary structure aspects, interactions between successive positions along the sequence, or terms related to the folding constraints, can all be combined in an additive manner in the inverse potential. In fact, the model need not even be formulated in terms of a Boltzmann distribution, as long as the parameters are fitted by ML, and the predictive power of the resulting models is evaluated in a systematic way. Altogether, this amounts to setting up a robust statistical framework helping us to understand how, and to what extent, the sequences of natural proteins are determined by protein structure.
We used Miyazawa and Jernigan's definition of contacts : each residue is represented by the center of its side chain atom positions; the positions of Cα atoms are used for glycine. Residues whose centers are closer than 6.5Å are defined to be in contact. The accessible surface of a residue is defined as the atomic accessible area when a probe of the radius of a molecule of water is rolled around the Van der Waal's surface of the protein . We used the program Naccess  to make this calculation. When treating PDB files with multiple chains, solvent accessibility was calculated taking into account all molecules in the structure. The accessibility classes (percentage relative to the accessibility in Ala-X-Ala fully extended tripeptide) were defined so as to generate D equal-sized subsets of sites. The complete definition of accessibility classes is available as supporting material [see Additional file 1].
Monte Carlo implementation
In order to calculate the derivative of ω in the gradient descent procedure, expectations with respect to p(S' | C, θ)in equation 33 are evaluated numerically. A sample drawn from p(S | C, θ) is obtained by a Gibbs sampling algorithm similar to that of Robinson et al. . The elementary cycle of our Gibbs sampler is as follows: for each p = 1..P, and for each i = 1..N P , each of the 20 amino acids is proposed at site i of protein p, by successively setting = a, for all a = 1..20; in each case, the energy change ΔG a induced by this point substitution is evaluated; then, is set to amino acid a with probability . After Q cycles of burnin, a series of h = 1..K EM cycles are performed, and after each cycle, the current sequence, S h , is recorded. Once the sample is obtained, the expectation (32) is evaluated as
and the derivative of ω with respect to θ follows immediately.
The overall gradient descent procedure runs as follows: we start from a random potential θ0 and a random set of sequences, and perform the following iterative scheme:
perform Q Gibbs cycles for the burnin, and k EM additional cycles for the sampling itself. Keep the final sequences as the starting point of the next cycle.
update θ by gradient descent, based on the estimate of the gradient obtained over the sample:
where . is a scalar product, and δθ is a step-vector. In practice, the coefficients of δθ are tuned empirically, allowing three degrees of freedom, for the α, the ε, and the μ component of the potential respectively.
As a stopping rule, we monitor the evolution of ω(θ) itself, which we evaluate every 100 steps by a numerical procedure (see below), and stop when ω(θ) has stabilized. In practice, we used Q = 100 and k EM = 100. At first sight, it would seem that a larger number of points k EM would be needed to get a precise expectation, but in the present case one can rely on the self-averaging of the derivatives across the 100,000 sites of the database.
The difficult part in estimating the likelihood (or equivalently ω(θ)), for a given value of θ, is to obtain an evaluation of ln Y. We do this by thermodynamic integration, or path sampling [51, 52], using the quasi-static method which we developed previously .
First, for 0 ≤ β ≤ 1, we define
The associated probability distribution is:
What we are looking for is In Y1. As for In Y0, it factors out, and can be computed directly:
We can thus equivalently evaluate the difference ln Y1 – ln Y0. To do this, we rely on the following identity:
where 〈·〉 β is the expectation over p β (s' | c, θ).
In practice, the method consists in first equilibrating the Gibbs sampler at β = 0, and then, performing a series of K Th + 1 cycles, where at each step, the value of β is increased by a small amount δβ = 1/K Th . The successive values of obtained during this quasi-static sampling scheme are recorded, and their average is our estimate of ln Y1 – ln Y0:
Note that these developments are for one protein, but the generalization over the database is straightforward.
In the conditions of the present work, K Th = 1, 000 is sufficient to obtain an estimate of ln Y1 – ln Y0 with an error less than one natural unit of logarithm.
We measured the fit of each model using cross-validation (CV): the potentials optimized on a first data set, i.e. the learning set, (θ L ) are applied on the second data set (the test set), and the log-likelihood is directly taken as a measure of fit. More precisely, for each model M,
CV M = -ln p(S T | C T , θ L , M), (47)
where S T and C T are the sequences and structures of the test set. The difference with the CV score obtained for the flat potential (μ) is reported: ΔCV = CV μ – CV M .
Sequence sampling: site-specific profiles
Once an optimal value of θ is obtained, sequences compatible with a given conformation can be sampled from p(s | c, ) by Gibbs sampling, and then further investigated. For instance, the frequency of each of the 20 amino acids (a) at each position (i) can be computed (q i (a)), yielding a vector of site-specific marginal profiles, graphically displayed as sequence logos . Alternatively, leave-one-out profiles can be obtained by computing the probability of each of the 20 amino-acids at each site of the test sequence, given the potential and the native sequence at all other positions:
p(s i = a | s j , j ≠ i, θ). (48)
We measured the amount of information displayed by the profiles using the site-specific Shannon entropy:
We compared both marginal and leave-one-out profiles to the empirical profiles, i.e. profiles displayed by natural sequences. We generated these empirical profiles from multiple sequence alignments obtained from the ConSurf-HSSP database .
Sequence sampling: design specificity
As a test for specificity, designed sequences were submitted to a fold recognition experiment, using the fold recognition program THREADER . In THREADER, the compatibility of a sequence s for a given structure c is measured by the Z-score:
where 〈E(S, C)〉 is the average of the THREADER statistical potential over all conformations of the decoy set, and σ is the corresponding standard deviation.
We randomly chose 70 structures of sizes ranging from 100 to 300 residues from the default THREADER dataset [see Additional file 8]. Structures whose native sequences produced a Z-score < 3 were discarded for the analysis. For each structure, c, we sampled 20 sequences from p(s | c, ) by Gibbs sampling. These designed sequences were then submitted to THREADER , and their specificity for the target structure c was measured by the ranking of c among all other structures, sorted by increasing Z-score.
A subset of 120 among the 1,200 sequences generated with the combined (ε + α14ac + μ) potential (3–5 sequences for 23 distinct conformations, chosen at random; [see Additional file 8]) were also submitted to another fold recognition program, LOOPP , and the presence of the native conformation c as the first hit or in the first 10 hits was recorded.
We used proteins culled from the entire PDB according to structure quality (resolution better than 2.0 Å) and with less than 25% of mutual sequence identity . Two subsets of approximately equal size were obtained by partitioning the proteins randomly: DS1, 449 proteins, 100,077 sites, and DS2, 465 proteins, 99,894 sites. The final list of proteins is available as supporting material [see Additional file 2] [see Additional file 3].
Drexler KE: Molecular engineering: an approach to the development of general capabilities for molecular manipulation. Proc Natl Acad Sci USA 1981, 78: 5275–5278. 10.1073/pnas.78.9.5275
Pabo C: Molecular technology: designing proteins and peptides. Nature 1983, 301: 200. 10.1038/301200a0
Ponders JW, Richards FM: Tertiary templates for proteins. Use of packing criteria in the enumeration of allowed sequences for different structural classes. J Mol Biol 1987, 193: 775–791. 10.1016/0022-2836(87)90358-5
Seno F, Vendruscolo M, Maritan A, Banavar JR: Optimal protein design procedures. Phys Rev Lett 1996, 77: 1901–1904. 10.1103/PhysRevLett.77.1901
Abkevich VI, Gutin AM, Shakhnovich EI: Improved design of stable and fast-folding model proteins. Fold Des 1996, 1: 221–230. 10.1016/S1359-0278(96)00033-8
Hellinga HW, Richards FM: Optimal sequence selection in proteins of known structure by simulated evolution. Proc Natl Acad Sci USA 1994, 91: 5803–5807. 10.1073/pnas.91.13.5803
Parisi G, Echave J: Structural constraints and emergence of sequence patterns in protein evolution. Mol Biol Evol 2001, 18: 750–756.
Bastolla U, Porto M, Roman HE, Vendruscolo M: Lack of self-averaging in neutral evolution of proteins. Phys Rev Lett 2002., 89:
Bastolla U, Porto M, Roman HE, Vendruscolo M: Connectivity of neutral networks, overdispersion, and structural conservation in protein evolution. J Mol Evol 2003, 56: 243–254. 10.1007/s00239-002-2350-0
Robinson DM, Jones DT, Kishino H, Goldman N, Thorne JL: Protein evolution with dependence among codons due to tertiary structure. Mol Biol Evol 2003, 20: 1692–1704. 10.1093/molbev/msg184
Rodrigue N, Lartillot N, Bryant D, Philippe H: Site interdependence attributed to tertiary structure in amino acid sequence evolution. Gene 2005, 347: 207–217. 10.1016/j.gene.2004.12.011
Park S, Yang X, Saven JG: Advances in computational protein design. Curr Opin Struct Biol 2004, 14: 487–494. 10.1016/j.sbi.2004.06.002
Wernisch L, Hery S, Wodak SJ: Automatic protein design with all atom force-fields by exact and heuristic optimization. J Mol Biol 2000, 301: 713–736. 10.1006/jmbi.2000.3984
Larson SM, England JL, Desjarlais JR, Pande VS: Thoroughly sampling sequence space: large-scale protein design of structural ensembles. Protein Sci 2002, 11: 2804–2813. 10.1110/ps.0203902
Dahiyat BI, Sarisky CA, Mayo SL: De novo protein design: towards fully automated sequence selection. J Mol Biol 1997, 273: 789–796. 10.1006/jmbi.1997.1341
Jaramillo A, Wernisch L, Héry S, Wodak SJ: Folding free energy function selects native-like protein sequences in the core but not on the surface. Proc Natl Acad Sci USA 2002, 99: 13554–13559. 10.1073/pnas.212068599
Miyazawa S, Jernigan RL: Estimation of effective interresidue contact energies from protein crystal structures: quasi-chemical approximation. Macromolecules 1985, 18: 534–552. 10.1021/ma00145a039
Sippl MJ: Boltzmann's principle, knowledge-based mean fields and protein folding. An approach to the computational determination of protein structures. J Comput Aided Mol Des 1993, 7: 473–501. 10.1007/BF02337562
Godzik A, Kolinski A, Skolnick J: Are proteins ideal mixtures of amino acids? Analysis of energy parameter sets. Protein Sci 1995, 4: 2107–2117.
Solis AD, Rackovsky S: Improvement of statistical potentials and threading score functions using information maximization. Proteins 2006, 62: 892–908. 10.1002/prot.20501
Hendlich M, Lackner P, Weitckus S, Floeckner H, Froschauer R, Gottsbacher K, Casari G, Sippl MJ: Identification of native protein folds amongst a large number of incorrect models. The calculation of low energy conformations from potentials of mean force. J Mol Biol 1990, 216: 167–180. 10.1016/S0022-2836(05)80068-3
Maiorov V, Crippen G: Contact potential that recognizes the correct folding of globular proteins. J Mol Biol 1992, 227: 876–888. 10.1016/0022-2836(92)90228-C
Mirny LA, Shakhnovich EI: How to derive a protein folding potential? A new approach to an old problem. J Mol Biol 1996, 264: 1164–1179. 10.1006/jmbi.1996.0704
Bastolla U, Farwer J, Knapp EW, Vendruscolo M: How to guarantee optimal stability for most representative structures in the protein data bank. Proteins 2001, 44: 79–96. 10.1002/prot.1075
Shakhnovich EI, Gutin AM: Engineering of stable and fast-folding sequences of model proteins. Proc Natl Acad Sci USA 1993, 90: 7195–7199. 10.1073/pnas.90.15.7195
Kurosky T, Deutsch JM: Design of copolymeric material. J Phys A Math Gen 1995, 27: L387-L393. 10.1088/0305-4470/28/14/003
Deutsch JM, Kurosky T: New algorithm for protein design. Phys Rev Lett 1996, 76: 323–326. 10.1103/PhysRevLett.76.323
Seno F, Micheletti C, Maritan A, Banavar JR: Variational approach to protein design and extraction of interaction potentials. Phys Rev Lett 1998, 81: 2172–2175. 10.1103/PhysRevLett.81.2172
Micheletti C, Seno F, Maritan A, Banavar J: Design of proteins with hydrophobic and polar amino acids. Proteins 1998, 32: 80–87. 10.1002/(SICI)1097-0134(19980701)32:1<80::AID-PROT9>3.0.CO;2-I
Banavar J, Cieplak M, Maritan A, Nadig G, Seno F, Vishveshwara S: Structure-based design of model proteins. Proteins 1998, 31: 10–20. 10.1002/(SICI)1097-0134(19980401)31:1<10::AID-PROT2>3.0.CO;2-L
Rossi A, Maritan A, Micheletti C: A novel iterative strategy for protein design. J Chem Phys 2000, 112: 2050–2055. 10.1063/1.480766
Rossi A, Micheletti C, Seno F, Maritan A: A self-consistent knowledge-based approach to protein design. Biophys J 2001, 80: 480–490.
Chiu TL, Goldstein RA: Optimizing potentials for the inverse protein folding problem. Protein Eng 1998, 11: 749–752. 10.1093/protein/11.9.749
Wald A: Note on the consistency of maximumm likelihood. Ann Math Stat 1949, 20: 595–601.
Sun S, Brem R, Chan R, Dill K: Designing amino acid sequences to fold with good hydrophobic cores. Protein Eng 1995, 8: 1205–1213.
Dempster A, Laird N, Rubin D: Maximum likelihood from incomplete data via the EM algorithm. J R Stat Soc B 1977, 39: 1–38.
Jones DT, Taylor WR, Thornton JM: A new approach to protein fold recognition. Nature 1992, 358: 86–89. 10.1038/358086a0
Meller J, Elber R: Linear optimization and a double statistical filter for protein threading protocols. Proteins 2001, 45: 241–261. 10.1002/prot.1145
Thomas PD, Dill KA: An iterative method for extracting energy-like quantities from protein structures. Proc Natl Acad Sci USA 1996, 93: 11628–11633. 10.1073/pnas.93.21.11628
Kono H, Saven JG: Statistical theory for protein combinatorial libraries. Packing interactions, backbone flexibility, and the sequence variability of a main-chain structure. J Mol Biol 2001, 306: 607–628. 10.1006/jmbi.2000.4422
Thomas PD, Dill KA: Statistical potentials extracted from protein structures: how accurate are they? J Mol Biol 1996, 257: 457–469. 10.1006/jmbi.1996.0175
Skolnick J, Jaroszewski L, Kolinski A, Godzik A: Derivation and testing of pair potentials for protein folding. When is the quasi-chemical approximation correct? Protein Sci 1997, 6: 676–688.
Tiana G, Colombo M, Provasi D, Broglia RA: Deriving amino acid contact potentials from their frequencies of occurrence in proteins: a lattice model study. J Phys Condens Matter 2004, 16: 2551–2564. 10.1088/0953-8984/16/15/007
Tobi D, Elber R: Distance-dependent, pair potential for protein folding: Results from linear optimization. Proteins 2000, 41: 40–46. 10.1002/1097-0134(20001001)41:1<40::AID-PROT70>3.0.CO;2-U
Vendruscolo M, Najmanovich R, Domany E: Can a pairwise contact potential stabilize native protein folds agaionst decoys obtained by threading? Proteins 2000, 38: 134–148. 10.1002/(SICI)1097-0134(20000201)38:2<134::AID-PROT3>3.0.CO;2-A
Koehl P, Levitt M: De novo protein design. I. In search of stability and specificity. J Mol Biol 1999, 293: 1161–1181. 10.1006/jmbi.1999.3211
Sippl MJ: Calculation of conformational ensembles from potentials of mean force. An approach to the knowledge-based prediction of local structures in globular proteins. J Mol Biol 1990, 213: 859–883. 10.1016/S0022-2836(05)80269-4
Betancourt MR, Skolnik J: Local propensities and statiscal potentials of backbone dihedral angles in proteins. J Mol Biol 2004, 342: 635–649. 10.1016/j.jmb.2004.06.091
Lee B, Richards M: The interpretation of protein structures: Estimation of static accessibility. J Mol Biol 1971, 55: 379–400. 10.1016/0022-2836(71)90324-X
Hubbard SJ, Thornton JM: Naccess. Depart of Biochem and Molec Biol University College London 1993.
Ogata Y: A Monte Carlo method for high dimensional integration. Numerische Mathematik 1989, 55: 137–157. 10.1007/BF01406511
Gelman A: Simulating normalizing constants: from importance sampling to bridge sampling to path sampling. Stat Sci 1998, 13: 163–185. 10.1214/ss/1028905934
Lartillot N, Philippe H: Computing Bayes factors using thermodynamic integration. Syst Biol 2006, in press.
Schneider TD, Stephens RM: Sequence Logos: a new way to display consensus sequences. Nucleic Acid Res 1990, 18: 6097–6100.
Glaser F, Rosenberg Y, Kessel A, Pupko T, Ben-Tal N: The ConSurf-HSSP Database: The Mapping of Evolutionary Conservation Among Homologs Onto PDB Structures. Proteins 2005, 58: 610–617. 10.1002/prot.20305
Wang G, Dunbrack RLJ: PISCES: a protein sequence culling server. Bioinformatics 2003, 19: 1589–1591. 10.1093/bioinformatics/btg224
Laskowski RA, Chistyakov VV, M TJ: PDBsum more: new summaries and analyses of the known 3D structures of proteins and nucleic acids. Nucleic Acids Res 2005, 33: D266-D268. 10.1093/nar/gki001
Authors are grateful to Thomas Simonson, Pierre Tufféry, Laurent Chiche, Jérome Gracy and Gertraud Burger, for their critical comments on the manuscript and useful discussions. This work was financially supported in part by the "60ème comission franco-québécoise de coopération scientifique". CLK was supported by NSERC, CIHR and the Université de Montréal; NR was supported by a bioinformatics grant from Genome Québec; HP by the Canada Research Chair Program and the Université de Montréal; CB and NL were funded by the french Centre National de la Recherche Scientifique, through the ACI-IMPBIO Model-Phylo funding program.
CLK participated in the implementation of the methods, performed the run of all the experiments, and co-wrote the manuscript. NR participated in the implementation of the methods, extensively supervised CLK and CB, and made contributions to the drafting of the manuscript. CB participated in the initial implementation of the methods. HP contributed to the drafting of the manuscript and the coordination of the project. NL set up the theoretical framework, co-wrote the manuscript, participated in the implementation of the methods and directed the overall project. All authors read and approved the final manuscript.
Electronic supplementary material
Authors’ original submitted files for images
About this article
Cite this article
Kleinman, C.L., Rodrigue, N., Bonnard, C. et al. A maximum likelihood framework for protein design. BMC Bioinformatics 7, 326 (2006). https://doi.org/10.1186/1471-2105-7-326
- Statistical Potential
- Solvent Accessibility
- Native Sequence
- Fold Recognition
- Thermodynamic Integration