 Research article
 Open Access
 Published:
An adaptive bin framework search method for a betasheet protein homopolymer model
BMC Bioinformaticsvolume 8, Article number: 136 (2007)
Abstract
Background
The problem of protein structure prediction consists of predicting the functional or native structure of a protein given its linear sequence of amino acids. This problem has played a prominent role in the fields of biomolecular physics and algorithm design for over 50 years. Additionally, its importance increases continually as a result of an exponential growth over time in the number of known protein sequences in contrast to a linear increase in the number of determined structures. Our work focuses on the problem of searching an exponentially large space of possible conformations as efficiently as possible, with the goal of finding a global optimum with respect to a given energy function. This problem plays an important role in the analysis of systems with complex search landscapes, and particularly in the context of ab initio protein structure prediction.
Results
In this work, we introduce a novel approach for solving this conformation search problem based on the use of a bin framework for adaptively storing and retrieving promising locally optimal solutions. Our approach provides a rich and general framework within which a broad range of adaptive or reactive search strategies can be realized. Here, we introduce adaptive mechanisms for choosing which conformations should be stored, based on the set of conformations already stored in memory, and for biasing choices when retrieving conformations from memory in order to overcome search stagnation.
Conclusion
We show that our bin framework combined with a widely used optimization method, Monte Carlo search, achieves significantly better performance than stateoftheart generalized ensemble methods for a wellknown proteinlike homopolymer model on the facecentered cubic lattice.
Background
Considering the close connection between the function of proteins and their threedimensional (tertiary) structure, there are many reasons for studying protein folding; these include the desire to predict protein function based on sequence data (via tertiary structure prediction), to better understand a number of diseases that are directly caused by protein misfolding, aggregation and fibrillogenesis (some of which include Alzheimer's, Huntington's and prion disease as well as cystic fibrosis [1]), and to design proteins with desired structure and function. Since experimental methods (Xray crystallography and NMR) for protein structure determination are highly labour intensive and require purification and, in the case of Xray crystallography, crystallization of proteins, computational methods for predicting protein structure from sequence are very attractive.
The ab initio protein folding problem is the problem of predicting the tertiary structure (the native state) of a protein from its amino acid sequence by minimizing a given energy function. Even for simple models that discretize conformations on a lattice (grid), this optimization problem is $\mathcal{N}$Phard [2, 3]. Its difficulty stems from the fact that the space of possible conformations is vast, and the search landscapes induced by the given energy function are very complex. One of the most prominent and successful approaches for solving this and many other $\mathcal{N}$Phard optimization problems is known as stochastic local search (SLS) [4]. Applied to protein folding problems, SLS methods attempt to find native states by iteratively performing small conformational modifications guided by the given energy function; randomized decisions are used to avoid getting trapped in local minima of the given energy landscape. Monte Carlo algorithms, which are widely used in protein structure prediction, are a prominent special case of SLS methods.
The performance of stochastic local search algorithms is critically dependent on the properties of the search landscape encountered, such as the number and distribution of local minima, the overall landscape ruggedness (measured, for example, using autocorrelation or fitnessdistance analysis), and the basin structure. Therefore, search strategies that can extract important features of the landscape and adapt the search accordingly are among the most effective tools for solving optimization problems with complex search landscapes. As evident from an analysis of the literature describing search methods for ab initio protein folding presented in the related work section, such adaptive search strategies have not been widely studied for this problem.
In this work, we introduce a novel adaptive SLS method that is based on a system of bins for storing a diverse set of conformations (candidate solutions) in memory. The general idea behind our approach is to store promising conformations encountered during the search for later use in situations when search stagnation is detected. These conformations are pooled into a number of bins according to energy and diversity criteria. The storage and retrieval mechanisms used in this context adaptively control the behaviour of a subsidiary local search procedure, such as canonical Monte Carlo search, and are shown to greatly improve its performance.
The remainder of this paper is structured as follows: First, we provide background information and discuss related work on protein folding problems and algorithms. Next, we introduce the bin framework and describe a Monte Carlo algorithm that utilizes this bin framework adaptively. We then compare the performance of our algorithm with that of other prominent methods from the literature, followed by a discussion of how the behavior and performance of our algorithm is influenced by its parameters. Finally, we explain how the bin framework introduced in this work relates to other search methods known from the literature, summarize our findings, draw some general conclusions, and indicate directions for future work. In the "Methods" section, we describe the facecentered cubic (FCC) lattice model and βsheet energy potential used in our computational analysis and provides details of the experimental protocols used in the empirical analysis of our algorithm.
Related work
To address the ab initio protein folding problem, the following three issues need to be considered: (1) the model used for the representation of protein structure (which may have implications on prediction accuracy); (2) the energy potential function (which ideally should be able to discriminate between native and nonnative conformations); and (3) the method used for searching through the space of possible conformations (which should be able to find optimal conformations as efficiently as possible). In this section, we discuss the choice of protein representation and energy function made in this work; we also provide a brief overview of the bestperforming search methods for ab initio protein folding known from the literature.
To facilitate ab initio protein structure prediction by means of more efficient methods for searching in the space of protein conformations, various reduced models of protein structure have been introduced by biochemists and physicists. These reduced models fall into two major classes: lattice and offlattice models. The primary reason for choosing offlattice models over lattice models is to obtain better geometrical accuracy. Despite the biases introduced by the lattice models, namely a somewhat restricted ability to accurately represent secondary structure and backbone conformation, lattice models still retain essential properties of the system [5–7] and offer a number of computational advantages; these advantages include fast energy computation, easiness of testing selfavoidance and the ability to use precomputed tables of moves, all of which help to compute search steps efficiently.
An important representative of lattice models is the FaceCentered Cubic (FCC) lattice, which underlies the crystalline structure of most metals. Even though in the context of ab initio protein structure prediction, the simpler square and cubic lattices are the most widely studied models in the literature, the FCC lattice captures real protein conformations with higher accuracy (coordinate C_{ α }root mean square deviation below 2 Å) [7] while still being representationally rather simple (it requires only 12 basis vectors). It has also been shown that local packing of amino acids in real proteins closely fits a distorted FCC lattice [8], and that the FCC model allows a reasonable description of secondary structure elements; furthermore, it can represent geometrically accurate hydrogen bonding [6]. As a result, the FCC lattice is considered the best overall choice among the simpler regular lattice models [6]. A detailed description of the FCC lattice is provided in the "Methods" Section.
As an energy function to be used in conjunction with the FCC lattice model we chose the βsheet potential [9, 10]. This choice was motivated by the fact that there are no universally used energy functions for ab initio protein structure prediction; at the same time, the βsheet potential has been used relatively widely in the literature for the empirical evaluation of the bestperforming protein structure prediction methods discussed later in this section. This enables us to compare our new approach against a relatively wide range of other conformation search methods known from the literature. Furthermore, this βsheet potential exhibits characteristics of more complicated energy functions used for offlattice models [9, 11] (particularly, cooperative allornone folding transition, characteristic interplay between short and longinteractions and secondary structure propensity). Finally, the problem of folding of βsheet proteins is particularly important, since the accuracy of protein structure prediction methods for βsheet proteins is the lowest among all structural classes of proteins [12]. The βsheet potential used in this work is described in detail in the "Methods" Section.
There are a number of search methods applicable to the protein folding problem that can be used in conjunction with reduced complexity models and simplified potentials to perform a broad search through lowresolution structures. The most widely used methods include Metropolis Monte Carlo (MC) search [13–17], Genetic Algorithms [18–20] and Generalized Ensemble Methods [21–23], which include the currently bestperforming algorithms for ab initio protein structure prediction. This last class of algorithms is based on the observation that canonical Monte Carlo methods sample conformations according to Boltzmann probabilities. For typical distributions of states (i.e., protein conformations) over energy levels this means that very high and, more importantly, very low energy conformations are rarely sampled. Generalized Ensemble Monte Carlo Methods attempt to overcome this problem by striving to perform a random walk in energy space by computing the density of states, by sampling expanded range of temperatures or by computing other physical quantities affecting transitions between the states during search.
Currently, the bestperforming Generalized Ensemble Method for ab initio protein structure prediction is Replica Exchange Monte Carlo search (REMC) [23], also known as the multiple Markov Chain method or Parallel Tempering [22]. In REMC, a number of noninteracting copies (replicas) of the given protein sequence are folded independently and at different temperature settings of the underlying canonical Monte Carlo search. Every few steps, pairs of replicas are exchanged (i.e., the temperature settings of the MC search performed on them are swapped) with a probability that depends on the energies of the respective conformations (using Boltzmann weighting). While other Generalized Ensemble Methods, such as Multicanonical (MUCA) Monte Carlo (or Entropy Sampling Monte Carlo) search [9], maintain only one conformation at any given time, the number of replicas required in REMC increases as the square root of the number of degrees of freedom (which in its turn increases linearly with sequence length) [23].
Improvements of REMC introduced in the literature include hybrid approaches between REMC (for the weight factor determination) and MUCA, or Simulated Tempering production runs [22–24]. The ParallelHat Tempering (PHAT) Monte Carlo method utilizes an additional weight factor based on the histogram of energies sampled by each temperature replica [10] to achieve an exponential increase in the acceptance probabilities for high and lowenergy conformations, which increases the efficiency of the search process by allowing it to effectively overcome higher energy barriers and to explore a wider range of conformations for each replica.
The following algorithms have been implemented and empirically evaluated for the FCC lattice model with the βsheet energy function considered in this work: canonical Metropolis Monte Carlo search [9], MUCA [9], REMC [9] and PHAT [10]. Of these, PHAT is the bestperforming conformation search method for the ab initio prediction of protein structures on the FCC lattice using the βsheet potential.
Bin framework Monte Carlo search
The key idea behind our adaptive bin framework is to improve the effectiveness of a given search process, such as canonical Monte Carlo search, by making it adaptive and augmenting it with a mechanism for storing and retrieving promising conformations. This is achieved by using a series of bins each of which holds a set of conformations within a certain energy range and an adaptive strategy for restarting a given search process with a conformation retrieved from these bins when the search stagnates. By varying the search strategy according to a priori defined transition probabilities (which are dependent on the search progress), this approach leads to an algorithm that sacrifices an exact relationship with the canonical ensemble for search efficiency. This method effectively reduces the slow convergence, or quasiergodicity, in rugged energy landscapes; it is therefore very useful when the main interest is in finding global minima, rather than in obtaining other physical properties from canonical ensembles.
Conformations are added to the bins in a way that is aimed at maintaining a diverse set of lowenergy conformations. To achieve this we define energy and diversity thresholds for each bin, which are dynamically modified during the search process (it may be noted that this adaptive strategy is closely related to the concept of reactive search [25]).
With each bin i the following properties are associated:

The capacity of the bin, cap_{ i }, i.e., the maximal number of conformations that can be stored in the bin at any given time.

The current number of conformations stored in the bin, num_{ i }. These conformations themselves are stored in a list that is sorted according to energy, to facilitate efficient retrieval of lowenergy conformations.

The width of the bin's maximal energy range, ΔE_{ i }.

The bin's energy threshold, ${E}_{i}^{+}$. This is the highest energy that a conformation can reach and still be placed into bin i if the respective diversity threshold (described in the following) is satisfied.

The diversity threshold, which determines how different a conformation has to be from other conformations already stored at the same energy level in order for the new conformation to be accepted into the bin. The pairwise diversity of conformations is measured using a distance measure that depends on the protein model under consideration. Here, we use the normalized average Hamming distance, HD, between the βsheet energy sequence of a newly considered conformation c and all βsheet energy sequences for the set C' of all conformations with the same energy that are already in the bin, see Methods Section for details.
Furthermore, the overall bin framework has the following parameters:

The total number of bins, numBins.

The energy range of interest, ΔE. Together with the current estimate of the ground state energy, $\widehat{E}$, which is modified throughout the search to always represent the lowest energy encountered so far, ΔE controls the energy interval into which conformations must fall in order to be accepted into any bin. This range represents the estimate of the maximal barrier height that needs to be surmounted to reach lower energy conformations.

The temperature T_{ bin }, which controls the retrieval of conformations from bins.
The general bin framework search mechanism is outlined in Figure 1. Procedure initalizeBins is used to set all bin parameters to their initial values. Bins are numbered 1 ... numBins, and for every bin i, ΔE_{ i }and ${E}_{i}^{+}$ are always assigned such that ${E}_{i}^{+}<{E}_{i1}^{+}\Delta {E}_{i1}+1$, i.e., the energy intervals $[{E}_{i}^{+},{E}_{i}^{+}\Delta {E}_{i})$ for different bins never overlap.
Furthermore, the energy bounds are always assigned such that ${E}_{i}^{+}>{E}_{2}^{+}>\cdots >{E}_{numBins}^{+}$, i.e., bin 1 always has the highest energy interval, while bin numBins stores the lowest energy conformations. Procedure subSearchStep performs one step of a subsidiary search procedure (such as canonical MC search) on a given conformation c and returns the resulting conformation c'. This step involves a single proposal of the move from the given move set and its subsequent acceptance or rejection. The two procedures considerStoringInBin and retrieveFromBin control the storage of conformations in the bin system and the retrieval of previously stored conformations. Note that conformations will only be stored in a bin if they satisfy the corresponding energy and diversity thresholds. Storing a conformation may lead to adjustments of the energy thresholds for the corresponding bin or addition of a new bin (this will be described later in detail for the BINMC algorithm). Finally, a stagnation criterion is used to decide when to retrieve a conformation from the bin system in order to overcome search stagnation, and a termination condition is used to determine when the search process should terminate.
In the following, we will describe the specific instantiation of this framework on which the remainder of our study is focused: the BINMC algorithm.
In BINMC, for simplicity all bin capacities cap_{ i }are set to the same value, and this value is kept constant during the search. The same holds for the energy ranges ΔE_{ i }. Finally, for simplicity we also keep the number of bins, numBins, constant during the run of the algorithm. This number is determined by the interval of energies of interest and the energy window width used:
numBins = ⌈ΔE/ΔE_{ i }⌉. (1)
At the beginning of the search process, the energy threshold for each bin i is set to ${E}_{i}^{+}$ := (i  1)·ΔE_{ i }. Initially, the energy intervals of all bins form a partition of the interval [0, numBins·ΔE_{ i }), note that this interval is larger than or equal to the desired interval [0, ΔE); it is larger if ΔE does not divide evenly by ΔE_{ i }.
It should be noted that 0 is the highest energy possible under the model chosen in this work and all the energies are integer values; in the general case energy thresholds can be adjusted initially to store the highest energy conformations under the protein model considered. The bin energy bounds are adjusted during the search, as will be explained later.
The diversity thresholds for the bins, HD_{ i }, are determined based on the following formula:
HD_{ i }= HD_{ min }+ (numBins  i)·(HD_{ max } HD_{ min })/(numBins  1), (2)
where HD_{ min }and HD_{ max }are parameters of the algorithm that determine the diversity threshold of the lowest and highest energy bins, respectively. This choice is based on the experimental observation that fewer protein conformations exist for lower energies, and therefore, the set of of conformations to be found at low energy levels can be expected to be less diverse. (This is also consistent with the prevalent view that the energy landscapes encountered in ab initio protein structure prediction problems are funneled [26].) Figure 2 depicts some of the properties of bins and conformations in a bin and the overall relationship of conformations stored in the framework to the energy landscape.
BINMC uses a standard canonical Monte Carlo search procedure running at a constant inverse temperature β_{ MC }:= 1/(k_{ B }·T_{ MC }) (where k_{ B }denotes the Boltzmann constant), that controls the probability with which worsening search steps are accepted. Our canonical Monte Carlo procedure for the FCC lattice is based on the same search neighbourhood (move set) as used by Gront et al. [9] and by Zhang and Skolnick [10]. In each search step, either a doublebond move or a chainend move is attempted. A doublebond move involves the modification of two successive bond angles, whose position in the chain is chosen uniformly at random. Similarly, in a chainend move, the location of the first or the last residue is changed. For efficiency and speed, doublebond moves are precomputed in a table, as done by Gront et al. [9] and by Zhang & Skolnick [10]. The search proceeds in phases, each of which starts with two attempts at chainend moves (one on each end, in random order) followed by N/2 successive attempts at doublebond moves (each chosen uniformly at random, without repetition but allowing overlap with previously chosen doublebond moves) any number of which may be accepted. Procedure subSearchStep in Figure 1 performs a single step of this simple subsidiary search process by attempting one move, starting from conformation c and resulting in conformation c' (which, if the proposed move has not been accepted, is equal to c); the attempted moves are chosen such that the previously described phasing of chainend and doublebond moves is ensured.
After a new conformation has been accepted by the subsidiary MC procedure, it is considered for placement into a bin. If the new conformation c' has lower energy then any other conformation encountered so far, i.e., if E(c') <$\widehat{E}$, it is always accepted into the bin framework and the current estimate of the ground state, $\widehat{E}$, is updated. If E(c') falls outside the energy range currently represented by the bin framework, before accepting the new conformation, a new bin (or a number of bins if needed) is created and the first bin storing conformations with high energies (or a number of bins starting with the first one) is deleted as follows: (here we only describe addition of a single bin, addition of multiple bins is handled analogously): We add a new bin numBins + 1 and delete all conformations from bin 1 along with bin 1 itself. We also shift bin numbers by 1, such that bin 2 becomes bin 1 and bin numBins + 1 becomes bin numBins. The energy threshold ${E}_{numBins}^{+}$ for the newly added bin is set to ${E}_{numBins1}^{+}\Delta {E}_{numBins1}+1$. The diversity thresholds are not shifted with the bins, such that HD_{ max }and HD_{ min }remain associated with the first and the last bin, respectively.
If conformation E(c') falls within the energy interval of a bin i that is not yet filled to capacity (i.e., num_{ i }< cap_{ i }), and c' satisfies the diversity criterion for that bin – i.e., the Hamming distance between the conformation c' and other conformations c" with the same energy E should be larger or equal to HD_{ i }(see Methods Section for details) – c' is added to that bin, and num_{ i }is incremented by one. (Note that there is at most one such bin i, since bin energy intervals never overlap.) Finally, if E(c') falls within the energy interval of a bin i that is already filled to capacity (i.e., num_{ i }= cap_{ i }) and it satisfies the diversity criterion for that bin, c' is added to the content of bin i and the highest energy conformation is removed. At the same time, $\widehat{E}$ is set to the energy of the conformation that is currently the highest in the bin; as a consequence, conformations with energy above the updated $\widehat{E}$ will not be accepted into bin i in the future, and therefore, the energy ranges of bins i  1 and i are now may no longer be adjacent.
The stagnation detection mechanism used in BINMC is quite simple: the search is considered to be stagnated when no improvement on the lowest energy has been achieved for noImprRetrieve search steps, where noImprRetrieve is a parameter of the algorithm.
To retrieve a conformation from the bin system, BINMC uses a twophase procedure that first selects a bin and then chooses one of the conformations stored in that bin. In the first phase, the probability of selecting a bin i depends on the difference between its upper energy threshold $\widehat{E}$ and the best energy reached so far, and is proportional to:
where β_{ bin }= 1/(k_{ B }·T_{ bin }), k_{ B }denotes the Boltzmann constant, and T_{ bin }is BINMC's temperature parameter. In the second phase, the probability of choosing a conformation c from that bin is analogously proportional to:
In general, conformations could be chosen with or without replacement; here we limited ourselves to choosing conformation with replacement, since the same conformation can yield a different fold each time it is picked.
As in the stochastic tunneling approach [27], to lessen exponential decay of the probability function we used Boltzmannbased modified weights proportional to ${e}^{\beta (E\widehat{E})}$. This weighting preserves the location of all minima, but maps the entire energy space from $\widehat{E}$ to the maximum energy 0 onto the interval [0, 1]. The dynamic process following the Boltzmann distribution can therefore pass through energy barriers of an arbitrary height.
Overall, this retrieval procedure ensures that lower energy conformations are selected with higher probability, which is in accordance with the belief that the energy landscapes of real proteins are funneled. The search is terminated when the target energy level has been reached or when a userspecified number of steps has been executed.
Results
In this section, we present empirical performance results for BINMC as compared to the bestperforming algorithms known from the literature. MC, REMC and PHAT have been tested by their original authors on the homopolymers of length 32 and 64, but only results for the homopolymer of length 64 have been published [9, 10]. (We contacted D. Gront [9] and Y. Zhang [10], who commented that all of their algorithms were able to reach quite easily what they believed to be the global minimum for the polymer of length 32. However, they could not provide any information regarding the energy values or conformations reached in these experiments.) For the protein of length 64, Gront et al. believed the lowest energy they had reached, 374, to correspond to a ground state conformation; however, Zhang and Skolnick later reported a conformation with energy 387 for this system [10].
Analogous with the results of Gront et al. for MC and REMC [9] and of Zhang and Skolnick for PHAT [10], in Table 1 we provide results averaged over 10 independent runs of each algorithm. It should be noted that several details were not evident from the published descriptions of these algorithms. From personal communication with D. Gront, we learned that their MC procedure was run with temperatures from 2.75 to 1.25 [ε_{0}/k_{ B }]. Since we could not determine the exact annealing schedule used in their study, we chose a constant temperature of 1.25 for our MC procedure. Therefore, the results for the MCSA algorithm of Gront et al. may not be exactly comparable with that of our implementation of pure MC. Since the number of iterations performed within a given amount of time varies significantly based on implementation details, we used the average CPU times reported by Gront et al. [9] and by Zhang and Skolnick [10] as the cutoff time for our algorithms.
As seen from Table 1, our implementations of MC and REMC show comparable performance to that of MCSA and REMC by Gront et al. [9]. (As discussed in the table caption, differences in execution environments were taken into account.) However, our implementation of PHAT did not reach the performance reported in the literature [10]. Therefore, we contacted Y. Zhang to check unpublished details of their algorithm, but unfortunately, those details along with the precise information on the best conformation reported in their paper (with energy 387) were no longer available due to data loss. Our novel BINMC algorithm performs better than MC and REMC, and than our implementation of PHAT. We used the following set of parameters for the bin framework for the homopolymer of length 64: ΔE = 30 [ε_{0}], ΔE_{ i }= 5 [ε_{0}], T_{ MC }= 1.25 [ε_{0}/k_{ B }], T_{ bin }= 6.521 [ε_{0}/k_{ B }], binCapacity = 100, HD_{ MAX }= 0.8, HD_{ MIN }= 0.01, noImprRetrieve = 2 000 000 steps. These settings were determined in a series of experiments in which we studied the influence of different parameter settings; these will be further discussed in Discussion Section.
The lowest energy level for the homopolymer of length 64 reached by our BINMC algorithm is 391; this is lower than the energy of any conformation previously reported in the literature. Conformations with energy 391 were found in 2 out of 10 runs, each with a 100 CPU hour cutoff on our reference machine, after 47 and 55 CPU hours, respectively. One of the two resulting conformations is shown in Figure 3, the other is its exact mirror image. Conformations with energies of 389 and 388 (some of which are shown in the supplementary material, see Additional file 1), were found multiple times by BINMC within a CPU time cutoff of 10 hours on our reference machines.
To extend our comparison of the reimplemented methods from the literature (MC, REMC and PHAT) with BINMC, we tested these methods on instances of length 12, 24 and 32. Again, we performed 10 independent runs of each algorithm on every problem instance, measuring the mean as well as the standard deviation of the energy levels reached after a runtime of 1 CPU hour. For BINMC when applied to the homopolymers of length 12 and 24, we used the following parameter settings: ΔE = 20 [ε_{0}], ΔE_{ i }= 5 [ε_{0}], T_{ MC }= 1.25 [ε_{0}/k_{ B }], T_{ bin }= 4.344 [ε_{0}/k_{ B }], binCapacity = 100, HD_{ MAX }= 0.6, HD_{ MIN }= 0.01 and noImprRetrieve = 100 000 steps, whereas on the homopolymer of length 32, we set the parameters to: ΔE = 20 [ε_{0}], ΔE_{ i }= 5 [ε_{0}], T_{ MC }= 1.25 [ε_{0}/k_{ B }], T_{ bin }= 4.344 [ε_{0}/k_{ B }], binCapacity = 100, HD_{ MAX }= 0.6, HD_{ MIN }= 0.01 and noImprRetrieve = 1 000 000 steps. (These parameter settings are discussed in Discussion Section and in the supplementary material, see Additional file 1.)
As can be seen from our results presented in Table 2, all methods find what appears to be the lowest energy (39) for the homopolymer of length 12 in less than 1 CPU second on our reference machine. For the homopolymer of length 24, we are starting to see differences among the algorithms: BINMC slightly outperforms all other algorithms in terms of CPU time required for reaching an energy of 109, which we believe to be the global minimum for this problem instance. MC is the next best method in terms of performance, followed by PHAT and REMC. The performance results for REMC and PHAT are worse than for MC because this homopolymer is too short for the additional time invested in exchanges between replicas to be amortized. For the homopolymer of length 32, BINMC outperforms the other algorithms by obtaining lower average energy (and also finding lower energy states, for example with energy 161, more often), followed by PHAT, REMC and finally MC. We show minimum energy conformations for the polymers of length 12, 24, and 32 in Figure 4. These solutions appear to be unique in terms of shortrange and longrange energy values, since all of the conformations found by any of the algorithms we ran show the same short vs longrange energy interplay.
Additionally, to see to which extent the results reported in Tables 1 and 2 could be further improved, we carried out 10 long independent runs for the homopolymers of length 64 and 32, using a cutoff time of 10 CPU hours on our reference machine. The results of this experiment are shown in Table 3; clearly, BINMC outperforms our implementation of the stateoftheart REMC and PHAT algorithms as well as the canonical MC algorithm in terms of the solution quality reached in these long runs, with REMC ranking second, followed by PHAT and MC.
Next, we conducted a more thorough performance comparison of the algorithms based on runtime distributions (RTDs) measured on the homopolymers of length 32 and 64. The goal of this analysis was to analyze the variability between independent runs on the same problem instance, which is known to reflect the parallelization efficiency of an algorithm and can also reveal detrimental stagnation behavior [4]. In order to be able to perform 100 independent runs of each algorithm on both sequences within a reasonable overall computation time using our reference machine, we used suboptimal energy levels of 158 and 370 for N = 32 and N = 64, respectively. The resulting empirical RTDs are shown in Figure 5 in the form of the respective cumulative distribution functions. For the homopolymer of length 32, BINMC and MC outperform REMC and PHAT in terms of the time required to reach the suboptimal solution quality of 158. For the homopolymer of length 64 BINMC performs best in terms of the time required to reach the suboptimal solution quality of 370, followed by MC, REMC and PHAT.
As can be seen from the graphs in Figure 5, the RTDs of all four algorithms closely resemble exponential distributions. (Note that the cumulative distribution functions of all exponential distributions have exactly the same shape when shown in a semilog plot.) This indicates that for reaching the energy levels considered here, none of the algorithms stagnates and all of them can be parallelized efficiently by concurrently executing independent runs [4].
It may come as a surprise that MC performs better than REMC and PHAT. However, it is important to note that the RTDs reported in Figure 5 are for suboptimal qualities only. Since MC at a low temperature is "greedier" and does not run multiple chains at different temperatures nor attempts exchanges between them, it gets to suboptimal energies faster. After reaching them, however, it stagnates; this is reflected in the observation that solution quality does not improve when running MC for a long time (10 CPU hours or more on our 2.4 GHz reference machine) for the homopolymers of length 32 and 64, as shown in Table 3. We also investigated the scaling behaviour of REMC and BINMC with homopolymer length. We measured the median runtime to reach the global minimum over 20 runs for sequences of length 12, 24, and 32 amino acids (the sequence of length 64 was not used, since only BINMC reaches the lowest known energy). The resulting sets of three data points for each algorithm were each fitted with a line in a semilogarithmic plot (which corresponds to fitting an exponential function to the original data in a way that counteracts overfitting for large instance sizes). Based on this analysis, the median runtime required for finding (purportedly optimal) conformations appears to scale as 10^{0.34·N5.2}for REMC and as 10^{0.28·N4.7}for BINMC.
Finally, we inspected the distribution of energies sampled by each method for the long homopolymer of length 64, based on approximately 5 × 10^{9} conformations each. As seen in Figure 6, REMC and PHAT show typical energy distributions for each replica, as reported by Zhang and Skolnick [10]; as expected, in the case of PHAT, the probabilities of encountering low and high energies are elevated. Interesting differences are observed when examining the distribution of energies visited by MC and BINMC (see Figure 7): While our new algorithm samples energies according to the Boltzmann probability density function, the distributions are shifted towards lower energy values, reflecting the fact that BINMC tends to reach lowerenergy conformations more efficiently.
Discussion
Similar to the Modelbased Search (MBS) method [28] introduced for offlattice fragment insertion as used in the ROSETTA [15] algorithm, our bin framework stores promising candidate solutions for future reuse. However, unlike MBS, we developed and tested an adaptive diversification mechanism that varies based on the energy level considered and takes into account how different a conformation is from other conformations with the same energy. Additionally, the energy level of interest, which determines the highest energy conformations stored in the bin framework are allowed to have, and individual Hamming distance criteria used for each bin, are adapted according to the current estimate of the ground state energy. MBS does not have a mechanism comparable to our diversity criteria between stored conformations. In MBS, a number of elite conformations are stored, whose quality is measured using a score determined from their energies and the radius of the local minimum represented by them. (This radius is estimated from the distance to their nearest neighbours using root mean square deviation.) Another important distinction between BINMC and MBS is that in BINMC, the model (a pool of stored conformations) is updated during the search and influences the choice of a new candidate solution every noImprRetrieve steps, when search stagnation is detected. On the other hand, in MBS for the discrete offlattice model with structural fragment insertion as described by Brunette and Brock [28], the choice of a new candidate solution is influenced at every step by the pool of conformations stored. Thus, MBS exploration of the search space is only dependent on the conformations stored [28]. Therefore, regions that are pruned based on the model are eliminated and not explored any further. In contrast, the bin framework provides only a subsidiary mechanism for generating candidate solutions when search stagnation is detected, and does not completely eliminate unexplored regions of the search space. This is achieved by running a nonmodelbased search (canonical MC) for a sufficiently long time to allow it to explore other regions of the search space.
The bin framework sorts conformations into bins representing different energy levels to make sure that the model contains as many energy levels of interest as possible while still reducing the search space. This aspect of the search is conceptually related to histogrambased sampling and search methods such as Multicannonical algorithm (MUCA) [24] and Energy Landscape Paving (ELP) method [21]. It should be noted, however, that our bin framework is a nonparametric model of the search space that consists of a diverse set of promising candidate conformations stored in the bins. MUCA, on the other hand, and to some degree ELP, are based on a parametric model of sampling all energy levels with the same probability without emphasis on low energies. In addition, our bin framework uses Hamming distance criteria that are based on the energy level of each bin to ensure that the respective sets of conformations stored are diverse and capture the overall funnellike structure of the landscape.
Up to this point, we have focused on comparing BINMC with existing algorithms. We now turn our attention to the question how the performance of the BINMC algorithm depends on its parameters and the algorithm components controlled by them, and to the determination of good settings for these parameters. We conducted this investigation for the homopolymer of length 32, since reaching its best known energy level of 161 is challenging, but not too computationally expensive to preclude performing multiple successful runs for a large number of parameter settings. For each parameter configuration, we conducted 20 independent runs on our reference machine and recorded the time required in each of these to reach an energy of 161; from these runs, we then determined the average CPU time required to reach this target energy. To study the influence of the different parameters on algorithm performance, we varied one (or, in the case of closely related parameters, two) of them at a time, while keeping all other parameter values fixed. Unless indicated otherwise, the parameters kept fixed in these experiments were set to the following values: ΔE = 20 [ε_{0}], ΔE_{ i }= 5 [ε_{0}], T_{ MC }= 1.25 [ε_{0}/k_{ B }], T_{ bin }= 4.344 [ε_{0}/k_{ B }], binCapacity = 100, HD_{ MAX }= 0.6, HD_{ MIN }= 0.1, noImprRetrieve = 100 000 steps.
Here, we summarize the results of this study; details are provided in the supplementary material, see Additional file 1. The impact of the parameters on the performance of BINMC can be ranked as follows (in decreasing order):

1.
the number of nonimproving steps (over the best energy) that are performed before reaching into a bin and replacing the current conformation in the Monte Carlo run, noImprRetrieve (the optimal value seems to be around 1 000 000 steps which allows the underlying MC search to explore the current region of the landscape);

2.
the ratio between the width of the energy of interest considered for binning, ΔE, and the bin temperature, T_{ bin }(the ratio that results in at least 0.01 probability of retrieval of the highest energy conformations works well);

3.
the diversity criteria used during binning high and lowenergy conformations (the Hamming distance limits), values of HD_{ min }= 0.01 and HD_{ max }= 0.06 guarantee that we are selective enough when storing promising conformations at low and high energy levels correspondingly and result in the best performance;

4.
the width of the energy window considered by each bin (ΔE_{ i }= 5 seem to provide optimal discretization of the search space); and

5.
the capacity of bins, cap_{ i }(storing 100 conformations in each bin seems to work best, this provides required diversification during the search).
Furthermore, we made the following observations regarding parameter settings of our algorithm for the longer homopolymer of length 64:

1.
a higher value of noImprRetrieve should be used (2 000 000 was found to work well), which indicates that longer search times are required to effectively explore the neighbourhood of the current conformation before reaching into the bin framework and replacing it with another promising conformation;

2.
ΔE should be increased, which is consistent with the common belief that the barrier heights for longer homopolymers are higher, and T_{ bin }should be adjusted such that the ratio ΔE/T_{ bin }remains the same as for length 32;

3.
ΔE_{ i }should be increased to E_{ i }= 10, which suggests that a coarser search space discretization may be beneficial for longer homopolymers; note that the combination of increases in ΔE and ΔEi results in only a slight increase in the total number of bins, numBins;

4.
the same values for HD_{ min }, HD_{ max }and cap_{ i }can be used as in the case of length 32.
Finally, our empirical results presented in Additional file 1 show that the BINMC algorithm performs substantially better than the simple restart strategy and than the pure Monte Carlo on which it is based. We also determined that all components of our algorithm are important for its efficiency, see Additional file 1. For example, the average time required for finding purportedly optimal conformations increases significantly (at least threefold for the homopolymer of length 32) when increasing noImprRetrieve until bin framework retrieval is performed very infrequently, when decreasing ΔE or cap_{ i }or when increasing the diversity thresholds HD_{ min }and HD_{ max }, resulting in very few conformations being stored, or when reducing the number of bins (by varying ΔE_{ i }for fixed ΔE).
Conclusion
The bin framework introduced in this work is a general approach that can be used to augment existing conformation search methods in order to increase their ability to focus on promising regions of the search phase (intensification) and to effectively overcome stagnation in regions of suboptimal conformations (diversification). As shown by our computational experiments, even very simple instantiations of the general bin framework can result in highly effective search algorithms.
In particular, our novel Bin Framework Monte Carlo algorithm (a combination of the bin framework and a simple Monte Carlo search procedure) surpasses Replica Exchange Monte Carlo search and its heuristic variant, Parallelhat Tempering, in its ability to find (purportedly) minimum energy conformations for βsheet homopolymers on the FCC lattice. Furthermore, using our new algorithm we have improved the best known solution for the homopolymer of length 64 from 387 to 391.
In future work, we plan to consider more advanced adaptive bin framework strategies that control search diversification and intensification reactively during the search, based on observed features of the search landscape. Additionally, we are planning to generalize our bin framework to work on partial as well as complete conformations, producing an efficient generalized framework that combines two distinct search strategies. Finally, we would like to extend our bin framework to address other models of protein structure, such as the FCC model with a more complex energy function or other discrete and continuous offlattice models.
Given the results for βsheet homopolymers on the FCC lattice achieved in this work, we believe that further investigation of our adaptive bin framework and its application to other protein structure prediction problems holds much promise.
Methods
In this section, we provide a detailed description of the FCC lattice model, βsheet energy potential, and experimental analysis performed to compare our approach to the existing methods from the literature.
FaceCentered Cubic lattice model
In the FCC model, the polypeptide is restricted to a facecentered cubic lattice [11] that has 12 base set vectors:
v_{ base }= {e_{1}, e_{2}, ..., e_{12}}, (3)
where e_{1} = (1, 1, 0), e_{2} = (1, 1, 0), e_{3} = (1, 0, 1), e_{4} = (1, 0, 1), e_{5} = (0, 1, 1), e_{6} = (0, 1, 1), e_{7} = (0, 1, 1), e_{8} = (0, 1, 1), e_{9} = (1, 0, 1), e_{10} = (1, 0, 1), e_{11} = (1, 1, 0), e_{12} = (1, 1, 0).
A protein chain of N residues is described by N  1 vectors, where vector v_{ i }connects residues i and (i + 1). The 12 base vectors allow for the following valence angles between each pair of vectors: 60°, 90°, 120°, and 180° [9].
Beta sheet energy potential used with the FCC lattice
To model βsheet proteins and the stiffness of the polymer chain, the following definition of an extended, β type chain conformation was defined by Gront et al. [9]: three vectors are in an extended state if

1.
the angles between vectors v_{i1}and v_{ i }and between v_{ i }and v_{i+1}are greater than 90 degrees;

2.
the dot product v_{i1}·v_{i+1}is larger than 0, which means that the angle between vectors v_{i1}and v_{i+1}is less than 90 degrees.
The energy potential for this model is composed of two terms: the shortrange potential U_{i1, i, i+1}that depends on the three consecutive vectors in the chain (v_{i1}, v_{ i }, v_{i+1}) and mimics conformational propensity to form an extended set of βstrands:
and the longrange potential V_{i, j}for two nonbonded chain residues, defined as:
where r_{i, j}is the lattice distance between residues i and j.
For a chain of length N, the total energy is defined as:
where δ_{ ij }is the Kronecker delta (δ_{ ij }= 1 when i = j, and 0 otherwise) and the values of the force field parameters are defined as ε_{ A }:= 1.0 [ε_{0}] and ε_{ B }:= 4.0 [ε_{0}] (with ε_{0} := 1.0 representing one unit of interaction energy) to model a semiflexible polymer [9].
Hamming distance criteria for the FCC lattice
For the FCC βsheet energy potential the normalized average Hamming distance, HD, between the βsheet energy sequence of a newly considered conformation c and all βsheet energy sequences for the set C' of all conformations with the same energy that are already in the bin is calculated as follows.
where $C=\{{{c}^{\prime}}_{1},\mathrm{...},{{c}^{\prime}}_{M}\}$, U_{ i }(x) denotes the U_{i1, i, i+1}value for conformation x, and ε_{ B }= 4.0 [ε_{0}] represents the energy contribution of each βresidue [9]. The inner sum ranges from i = 2 to i = N  2, since the first residue and the two last residues can never be in the extended βstate.
Empirical analysis
BINMC has been implemented in C++ and compiled using gcc (version 3.3.3) for the Linux operating system. The same holds for our implementations of three published algorithms for the same protein models: simple Monte Carlo (MC), Replica Exchange Monte Carlo (REMC) and Parallelhat (PHAT) Monte Carlo search [9, 10]; we had to reimplement these because the respective codes could not be obtained from the authors. All experiments were performed on PCs with 2.4 GHz Pentium IV CPUs, 256 Kb cache, and 1 Mb RAM, running Redhat Linux (our reference machine). Their runtime was measured in terms of the CPU time required to reach (or exceed) a specified energy level.
For our performance analysis we used the FCC βhomopolymers of length 12, 24, 36, and 64. The algorithms were evaluated based on a number of independent runs on each homopolymer. In most experiments, each run was terminated after a fixed CPU time limit (cutoff time) had been reached. From the distribution of energy levels over 10 independent runs, we determined the average energy, median energy, 25 and 75percentiles as well as the lowest energy reached. To further evaluate the performance of our BINMC algorithm and the methods known from the literature, we followed the methodology of Hoos and Stützle [29] and analyzed runtime distributions (RTDs) of the algorithms, i.e., the (empirical) probability distributions over the runtime required to reach the lowest known (or, in some cases, certain suboptimal) energy level for the respective homopolymer based on 100 independent runs.
References
 1.
Notling B: Protein Folding Kinetics. 1993, Springer
 2.
Ngo JT, Marks J, Karplus M: Computational Complexity: Protein Structure Prediction and the Levinthal Paradox. Protein Engineering. 1992, 5: 313321. 10.1093/protein/5.4.313.
 3.
Unger R, Moult J: Finding the Lowest Free Energy Conformation of a Protein is a NPhard Problem: Proof and Implications. Bulletin of Mathematical Biology. 1993, 55: 11831198.
 4.
Hoos HH, Stützle T: Stochastic Local Search: Foundations and Applications. 2004, Morgan Kaufmann Publishers, Elsevier
 5.
Bonneau R, Baker D: Ab Initio Protein Structure Prediction: Progress and Prospects. Annual Review of Biophysics and Biomolecular Structure. 2001, 30: 173189. 10.1146/annurev.biophys.30.1.173.
 6.
Kolinski A, Skolnick J: Reduced Models of Proteins and their Applications. Polymer. 2004, 45: 511524. 10.1016/j.polymer.2003.10.064.
 7.
Park BH, Levitt M: The Complexity and Accuracy of Discrete State Models of Protein Structure. Journal of Molecular Biology. 1995, 249: 493507. 10.1006/jmbi.1995.0311.
 8.
Bagci Z, Jernigan RL, Bahar I: Residue coordination in Proteins Conforms to the Closest Packing of Spheres. Polymer. 2002, 43: 451459. 10.1016/S00323861(01)00427X.
 9.
Gront D, Kolinski A, Skolnick J: Comparison of Three Monte Carlo Conformational Search Strategies for a Proteinlike Homopolymer Model: Folding Thermodynamics and Identification of LowEnergy Structures. Journal of Chemical Physics. 2000, 13: 50655071. 10.1063/1.1289533.
 10.
Zhang Y, Skolnick J: ParallelHat Tempering: A Monte Carlo Search Scheme for the Identification of LowEnergy Structures. Journal of Chemical Physics. 2001, 115: 50275032. 10.1063/1.1396672.
 11.
Pokarowski P, Kolinski A, Skolnick J: A Minimal Physically Realistic ProteinLike Lattice Model: Designing and Energy Landscape that Ensures AllOrNone Folding to a Unique Native State. Biophysical Journal. 2003, 84: 15181526.
 12.
Aloy P, Stark A, Hadley C, Russel RB: Predictions without Templates: New Folds, Secondary Structure, and Contacts in CASP5. Proteins: Structure, Function, and Genetics. 2003, 53: 436456. 10.1002/prot.10546.
 13.
Jones DT: Predicting Novel Protein Folds by Using FRAGFOLD. Proteins: Structure, Function, and Genetics Suppl. 2001, 5: 127132. 10.1002/prot.1171.
 14.
Ortiz AR, Kolinski A, Rotkiewicz P, Ilkowski B, Skolnick J: Ab Initio Folding of Proteins Using Restrains Derived from Evolutionary Information. Proteins: Structure, Function, and Genetics Suppl. 1999, 3: 177185. 10.1002/(SICI)10970134(1999)37:3+<177::AIDPROT22>3.0.CO;2E.
 15.
Simons K, Kooperberg C, Huang E, Baker D: Assembly of Protein Tertiary Structures from Fragments with Similar Local Sequences Using Simulated Annealing and Bayesian Scoring Function. Journal of Molecular Biology. 1997, 268: 209225. 10.1006/jmbi.1997.0959.
 16.
Skolnick J, Kolinski A, Kihara D, Betancourt M, Rotkiewicz P, Boniecki M: Ab Initio Protein Structure Prediction via a Combination of Threading, Lattice Folding, Clustering, and Structure Refinement. Proteins: Structure, Function, and Genetics Suppl. 2001, 5: 149156. 10.1002/prot.1172.
 17.
Zhang Y, Kihara D, Skolnick J: Local Energy Landscape Flattening: Parallel Hyperbolic Monte Carlo Sampling of Protein Folding. Proteins: Structure, Function, and Genetics. 2002, 48: 192201. 10.1002/prot.10141.
 18.
Bowie JU, Eisenberg D: An Evolutionary Approach to Folding Small αhelical Proteins that Use Sequence Information and an Empirical Guiding Fitness Function. Proceedings of the National Academy of Sciences of the USA. 1994, 91: 44364440. 10.1073/pnas.91.10.4436.
 19.
Dandekar T, Argos P: Folding the Main Chain of Small Proteins with the Genetic Algorithm. Journal of Molecular Biology. 1994, 236: 844861. 10.1006/jmbi.1994.1193.
 20.
Pedersen JT, Moult J: Protein Folding Simulations with Genetic Algorithms and a Detailed Molecular Description. Journal of Molecular Biology. 1997, 269: 240259. 10.1006/jmbi.1997.1010.
 21.
Hansmann UHE: Protein Folding Simulations in a Deformed Energy Landscape. The European Physical Journal B. 1999, 12: 607611. 10.1007/s100510051044.
 22.
Mitsutake A, Sugita Y, Okamoto Y: GeneralizedEnsemble Algorithms for Molecular Simulations of Biopolymers. Biopolymers (Peptide Science). 2001, 60: 96123. 10.1002/10970282(2001)60:2<96::AIDBIP1007>3.0.CO;2F.
 23.
Sugita Y, Okamoto Y: ReplicaExchange Multicanonical Algorithm and Multicanonical ReplicaExchange Method for Simulating Systems with Rough Energy Landscape. Condensed Materials Archive. 2004, 0009119
 24.
Okamoto Y: GeneralizedEnsemble Algorithms: Enhanced Sampling Techniques for Monte Carlo and Molecular Dynamic Simulations. Journal of Molecular Graphics and Modeling. 2004, 22: 425439. 10.1016/j.jmgm.2003.12.009.
 25.
Battiti R, Tecchiolli G: The Reactive Tabu Search. ORSA Journal on Computing. 1994, 6 (2): 126140.
 26.
Plotkin SS, Onuchic JN: Understanding Protein Folding with Energy Landscape Theory Part I: Basic Concepts. Quarterly Reviews of Biophysics. 2002, 35: 111167. 10.1017/S0033583502003761.
 27.
Wenzel W, Hamacher K: Stochastic Tunneling Approach for Global Minimization of Complex Potential Energy Landscapes. Physical Review Letters. 1999, 82: 30033007. 10.1103/PhysRevLett.82.3003.
 28.
Brunette TJ, Brock O: Improving Protein Structure Prediction with Modelbased Search. Bioinformatics. 2005, 21: i66i74. 10.1093/bioinformatics/bti1029.
 29.
Hoos HH, Stützle T: On the Empirical Evaluation of Las Vegas Algorithms. Proceedings of the 14th Conference on Uncertainty in Artificial Intelligence. 1998, 238245.
Acknowledgements
This work has been supported by an NSERC Postgraduate Scholarship (PGSB) held by AS and by funding from the Mathematics of Information Technology and Complex Systems (MITACS) Network of Centers of Excellence held by HH.
Author information
Additional information
Authors' contributions
Both authors contributed to the development of ideas, design of experiments, analysis and interpretation of results, and the writing of the paper. AS implemented the proposed method and performed the computational experiments.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Received
Accepted
Published
DOI
Keywords
 Monte Carlo
 Protein Structure Prediction
 Monte Carlo Search
 Replica Exchange Monte Carlo
 Reference Machine