 Research article
 Open Access
 Published:
SiteSeek: Posttranslational modification analysis using adaptive localityeffective kernel methods and new profiles
BMC Bioinformatics volume 9, Article number: 272 (2008)
Abstract
Background
Posttranslational modifications have a substantial influence on the structure and functions of protein. Posttranslational phosphorylation is one of the most common modification that occur in intracellular proteins. Accurate prediction of protein phosphorylation sites is of great importance for the understanding of diverse cellular signalling processes in both the human body and in animals. In this study, we propose a new machine learning based protein phosphorylation site predictor, SiteSeek. SiteSeek is trained using a novel compact evolutionary and hydrophobicity profile to detect possible protein phosphorylation sites for a target sequence. The newly proposed method proves to be more accurate and exhibits a much stable predictive performance than currently existing phosphorylation site predictors.
Results
The performance of the proposed model was compared to nine existing different machine learning models and four widely known phosphorylation site predictors with the newly proposed PSBenchmark_1 dataset to contrast their accuracy, sensitivity, specificity and correlation coefficient. SiteSeek showed better predictive performance with 86.6% accuracy, 83.8% sensitivity, 92.5% specificity and 0.77 correlationcoefficient on the four main kinase families (CDK, CK2, PKA, and PKC).
Conclusion
Our newly proposed methods used in SiteSeek were shown to be useful for the identification of protein phosphorylation sites as it performed much better than widely known predictors on the newly built PSBenchmark_1 dataset.
Background
Posttranslational modifications are observed on almost all proteins analysed to date. During phosphorylation, a phosphate molecule is placed on another molecule resulting in the functional activation or inactivation of the receiving molecule. These modifications have a substantial influence on the structure and functions of protein. Phosphorylation at the serine, threonine and tyrosine residues by enzymes of the kinase and phosphatise superfamilies is one of the most frequent forms of posttranslational modifications in intracellular proteins. As phosphorylation has a significant impact on diverse cellular signalling processes, it is needed in the regulation of cell differentiation, as a trigger for the progression of the cell cycle and control of metabolism, transcription, apoptosis, cytoskeletal rearrangements [1–7] in animals. As importantly, the phosphorylation of protein is considered as being a key event in many signal transduction pathways of biological systems [8]. It is thus important for us to be able to accurately determine the phosphorylation state of proteins so as to better identify the state of a cell.
It has been widely reported in literature that a large number of human diseases are caused by a disruption of normal cellular phosphorylation events. For example, phosphorylated tyrosines are recognised by specialised binding domains on other proteins, and such interactions are used to initiate intracellular signalling pathways. Aberrant tyrosine phosphorylation is a hallmark of many types of cancer. Tyrosine phosphorylation also plays major roles in cellular physiology, and functional perturbation of proteintyrosine kinases and proteintyrosine phosphatases underlie many human diseases [9].
In order to determine phosphoproteins and individual phosphorylation sites, various experimental tools have been used. However, in vivo or in vitro identification of phosphorylation sites is labourintensive, timeconsuming and often limited to the availability and optimisation of enzymatic reactions [7, 8, 10]. Several largescale phosphoproteomic data using the massspectrometry approach have been collected and published [11–13] but are still not helpful in distinguishing the kinasespecific sites on substrates. For example, mass spectrometry methods have been shown to be disfavourable in the identification of phosphatemodified residues, leading to an underestimation of the extent of phosphorylation presents in vivo[14].
Due to the practical limitations and complexities of the previouslymentioned methods, many scientists have turned to computerbased methods. Computerbased methods can efficiently handle massive amounts of protein data, determine phosphoprotiens and identify individual phosphorylation sites from one dimensional atomic coordinates with high precision. Several computersimulated machine learning techniques such as Artificial Neural Networks (ANNs) and Support Vector Machines (SVMs) have been extensively used in various biological sequence analyses as well as phosphorylation site prediction. These methods are built based on the assumption that neighbouring residues to the phosphorylated site represents the main determinant for kinase specificity [10, 15].
Although a large number of machine learning based methods have proved to be effective in the prediction of phosphorylation sites, several important issues that could potentially degrade the performance of machine learning or statisticalbased methods have been largely ignored. The high dimensionality of protein sequence data not only causes a dynamic increase in computational complexity but also creates an overfitting/generalisation problem for nonparametric methods. With machine learning models, better generalisation and faster training (computationally efficient) can be achieved when they have fewer weights to be adjusted with fewer inputs.
This study aims to develop an accurate and stable phosphorylation site predictor. Our proposed model named, SiteSeek uses a semiparametric form of a stateoftheart machine learning model dubbed, Adaptive LocalityEffective Kernel Machine (AdaptiveLEKM). In addition, with the boosting algorithm, it adaptively combines the learners to find an optimised fit for the given phosphoprotiens. To efficiently capture suitable information from protein sequences, it uses a newly developed Compact Evolutionary and Hydrophobic Profile (CEHProfile) built on Position Specific Scoring Matrix (PSSM) and Simultaneously Axially and Radially Alignment Hydrophobicity (SARAH1) Hydrophobicity scale. In our experiments, the SiteSeek excels in efficiently processing high dimensional protein data with a more accurate and stable predictive performance than currently existing models. The novel feature of this study is the use of a new machine learning based semiparametric model, a newly developed profile (CEHProfile) and the use of unique training dataset (PSBenchmark_1) that contains experimentally verified phosphorylation sites manually extracted from major protein sequence databases and the literature.
Results
Our experiments consist of four consecutive steps. First, we demonstrate the usefulness of our proposed CEHProfile by comparing its prediction accuracy with four other wellknown amino acids encoding methods. Second, the predictive performance of our proposed machine learning model, AdaptiveLEKM, specially designed for the high dimensional problem of protein sequence data is compared with other nine contemporary machine learning models for prediction accuracy, sensitivity, specificity, correlationcoefficient, type I and type II errors on newly built PSBenchmark_1 dataset. Next, we analyse the result from SiteSeek which uses the novel AdaptiveLEKM and the CEHProfile when tested on four main kinase groups and four main kinase families. We then compare those results with the consensus results from literature. Lastly, the predictive performance of SiteSeek is directly compared with thee of the most widely known contemporary phosphorylation site predictors on PSBenchmark_1 dataset.
Compact evolutionary and hydrophobicity profile
In order to prove the usefulness of CEHProfile, the suitability of the newly proposed profile to the prediction of phosphorylation sites is compared with other four most widely used encoding schemes (PSSM, SARAH1, Hydrophobicity Scale and Orthogonal Encoding). As shown in Table 1, the CEHProfile and CompactPSSM showed fairly improved performance for the prediction of phosphorylation sites. With CompactPSSM, we obtain 3% increased prediction accuracy than the widely used PSSM. Furthermore, CEHProfile which additionally uses hydrophobicity in the format of effective 5 bit binary representation (SARAH1 Scale) has a better accuracy at 5.1% – 6.5% than original PSSM. The sensitivity of CEHProfile has significantly increased by 5% from the original PSSM. It seems that with CEHProfile, we can combine the advantage of using both PSSM and SARAH1 profiles as SARAH1 provides a good model sensitivity level of 0.80.
This experiment based on our hypothesis proved that less discriminatory features that reside in original PSSM can be replaced by an effective hydrophobicity scale so that a new profile which contains compact evolutionary information as well as hydrophobic values can be created. As examined in Discussion, several limitations of existing PSSM and OE profiles can be effectively minimised by utilising the novel CEHProfile. Evidently, CEHProfile is more useful than widely known sequence profiles for protein phosphorylation site prediction.
Comparison of AdaptiveLEKM with other machine learning models
The predictive performance of AdaptiveLEKM was compared with nine other existing stateoftheart machine learning models such as General Regression Neural Network (GRNN), Radial Basis Neural Network (RBFN), MultiLayered Perceptron (MLP), kernel Nearest Neighbour (kNN), Decision Tree (J48), kernel Logistic Regression (KLR), and three different transductive SVMs, namely SVM^{light}, AdaBoostSVM and LocalityEffective Kernel Machine (LEKM). Table 2 shows the evaluation results of each model in terms of Accuracy (Ac), Sensitivity (Sn), Specificity (Sp), CorrelationCoefficient (Cc), Variance (Var) and Time on PSBenchmark_1 dataset.
As shown in Table 2, one of our models, LEKM successfully reached the best model stabilisation (Var: 0.020) with less computational requirements (Time: 22.421). However, one of the methods used in LEKM (semiparametric approximation) showed a slightly less accurate learning. Hence, we utilised AdaBoost algorithm for the fine tuning of the LEKM and it (AdaptiveLEKM) finally achieved the best accuracy with a fair level of model stableness and furthermore reduced complexity. In addition, AdaptiveLEKM achieved much better model robustness than other methods with the Cc of 0.39. Our methods used in AdaptiveLEKM, semiparametric approximation and adaptive tuning of the model using AdaBoost were confirmed to be more suitable in processing high dimensional protein data than other nonparametric models. It should be noted that the AdaBoost algorithm was also tested with the original SVM, but no significant improvement was observed (Table 2).
Figure 1 shows the comparison of prediction scores simulated by the AdaptiveLEKM and the original SVM on a protein chain (SwissProt Entry: O75553). The protein chain has 588 residues with two tyrosine and one serine sites at the residue 198, 220 and 524 respectively. As shown in Figure 1, SVM's signal at the site is generally 0.3 point with many fluctuating neighbouring signals making the site undistinguished. On the other hand, AdaptiveLEKM provides very clearer indications of the phosphorylation site at the residue 524 and its signal is generally stronger than that of other methods (0.37921 point). Thus, AdaptiveLEKM offers an additional advantage over other machine learners with a clearer and stronger indication of site locations.
Predictive performance of SiteSeek on major kinase families and groups
Here we look at experimental results obtained by SiteSeek on four main kinase groups and four main kinase families in terms of Ac, Sn, Sp, Cc, Type I and Type II ERs. SiteSeek uses the AdaptiveLEKM and is trained with the novel CEHProfile. Table 3 compares the results of SiteSeek with the consensus results of the literature. In general, SiteSeek showed a 9% increase in prediction accuracy than the consensus results. As for model stableness, SiteSeek also achieved a fairly low level of average variance in four evaluation measures (0.041, 0.074, 0.019, 0.083). The sensitivity of SiteSeek on CDK, PKA and PKC kinase families are distinguishably higher than the consensus results. This means that stable prediction capability of our model comes with effectively reducing the false negative values (Type I ER). Type I ER indicates experimentally verified unmodified sites that are predicted (incorrectly) to be modified.
As for the four main kinase groups (Table 4), the overall results are somewhat lower than those of kinase families. The best accuracy was obtained with the CMGC kinase group (Ac: 0.90) whereas the TK kinase group showed the lowest accuracy of 0.79. Compared to the experiments on kinase families, Cc values dropped significantly for 0.9 points. This result may reveal that computational models are more suitable to find stronger correlations among same kinase families than those of kinase groups.
Comparison with other contemporary phosphorylation site predictors
The predictive performance of our proposed predictor, SiteSeek which uses abovementioned AdaptiveLEKM and CEHProfile is compared with four contemporary phosphorylation site predictors on same testing dataset, PSBenchmark_1. DISPHOS [16] is trained on over 2000 nonredundant experimentally confirmed protein phosphorylation sites and uses positionspecific amino acid frequencies as well as disorder information to improve the discrimination between phosphorylation and nonphosphorylation sites. The prediction accuracy of DISPHOS reaches 81.3% for Serine, 74.8% for Threonine, and 79.0% for Tyrosine. However, DISPHOS provides little information about the corresponding protein kinases for the predicted phosphorylation sites. Scansite [17] identifies short protein motifs that are recognised by phosphorylation protein serine/threonine or tyrosine kinases. Each motif used in Scansite was constructed from a set of experimentally validated phosphorylation sites and was represented as a positionspecific scoring matrix. NetPhosK [18] is an enhanced version of NetPhos uses an artificial neural network and incorporates the functionality of providing protein kinases (PKs) information for 17 PKs. In the identification of protein kinase A (PKA) phosphorylation sites, NetPhosK achieves 100% sensitivity and 40% specificity. Finally, PredPhospho [19] uses support vector machine and demonstrates a superior predictive performance of 83–95% at the kinase family level, and 76–91% at the kinase group level.
Table 5 shows the comparison of our proposed model with the abovementioned four contemporary phosphorylation site predictors on the PSBenchmark_1 dataset. SiteSeek has shown the best overall accuracy amongst the stateoftheart predictors by reaching a prediction accuracy of 86.6%. Interestingly, the sensitivity of SiteSeek was far more superior to other predictors with 84.4%. In addition, its model robustness was again confirmed as Cc value showed 8.3 points higher than widely known Scansite. This result corresponds with the previous experimental results in that the newly proposed CEHProfile provides more useful information to the predictors for kinase family level prediction. In addition, the results give evidence that the semiparametric approach of AdaptiveLEKM which brings a more stable prediction is a better for phosphorylation site prediction than current existing methods. The prediction ability of the proposed model, which sustains its stability, was again proved with PSBenchmark_1 dataset.
Discussion
Over the past decades, many computational prediction algorithms have been developed for various proteomic studies. They have evolved from simple linear statistics to complex machine learners. However, the most significant breakthroughs were the incorporation of new biological information into an efficient prediction model and the development of new models which can efficiently exploit suitable information from its primary sequence. For example, the exploitation of evolutionary information that is available from protein families has brought significant improvements in the prediction of protein secondary structure (about 6–8%) [20–24].
Compared to protein structure predictions, a small number of studies to find suitable information/representations for phosphorylation site prediction have been reported. Like protein secondary structure prediction, most of the phosphorylation site prediction methods also use the evolutionary information in the format of PSSM (sequence profile) [16–18, 25]. The theory behind using the sequence profile is based on the fact that the sequence alignment of homologous proteins accords with their structural alignment and aligned residues usually have similar structures. Thus, the sequence profile can provide more information about structure than the single sequence to its learner.
Although the sequence profile provides more structural information, the structural information that resides in the sequence profile may not be a significant importance in the case of phosphorylation site prediction. It has been observed that approximately only ten neighbouring residues are major determinants of phosphorylation sites. Many models have been built on this observation and performed reasonably well with a number of specific kinases. However, the specificity determinants and rules remain elusive for a large number of protein kinases that display a number of substrates sharing little or no sequence similarity in the known phosphopeptides [10]. Furthermore, most databases searched by current alignment tools like PSIBLAST not only contain a number of nonphosphoprotiens, but also generate a large numbers of irrelevant hits from the protein databases [8].
Another commonly used amino acid encoding method is dubbed orthogonal encoding, also known as the binary representation or the distributed encoding. In orthogonal encoding, each letter can be represented by a twenty dimensional binary vector indicating the presence of a particular amino acid type [26]. The twenty standard amino acids are ordered one through twenty, and the i^{th}amino acid has the binary codeword of twenty bits with the i^{th}bit set to "1" and all others to "0"s, for i = 1, 2,..., 20. For example, 'Alanine' is expressed by 0000000000 00000000001, 'Cysteine' is encoded as 0000000000 0000000010 and so on. Among the first twenty units of the vector, each unit stands for one type of amino amid. In order to allow a window to extend over the N terminus and the C terminus, the 21st unit has to be added. A residue with window size n is encoded in 21*(2n+l) bits with the binary codewords of amino acids concatenated based on their order in the window [27]. It has been one of the most widely adopted methods as it does not introduce any artificial correlations between the amino acids.
However, orthogonal encoding has several widely known drawbacks. Firstly, the dimension of residue vector can increase rapidly as n increases, it may lead to large computational cost and model complexity (a typical input window of 13 residues requires 567 = (21*(2*13+1) input nodes and connecting weights), and recognition bias. Thus, it can cause poor performance of the classifiers [28–30]. Secondly, the use of Euclidean space has no theoretic foundation in biology or chemistry and hence might reduce the accuracy of a model. According to the numerical assignment in the distributed method, the distance between any two different amino acids is $\sqrt{2}$ and would conflict with reality in biology [29].
The encoding methods discussed above are employed by most wellknown protein structure predictors and are shown to be useful as they sufficiently contain information required for general protein structure prediction tasks. However, as phosphorylation site prediction does not only involve various chemical interactions but also is known as a nonstructural prediction task, the abovementioned encoding methods may not be suitable for this problem. In this study, we observed that widely known sequence profile contains irrelevant features which impede accurate recognition by the learner. We thus developed a new profile that replaces the prediction irrelevant features by hydrophobic information represented in SARAH1 scale. This method when used in CEHProfile showed reasonable improvement over existing profiles.
In the literature, the number of encoding schemes proposed for phosphorylation site prediction is far less than the ones for other proteomic applications. As discussed above, existing methods have shown several critical drawbacks for phosphorylation site prediction. Hence, we emphasise that researchers should devote their effort to seeking a suitable representation of amino acids for phosphorylation site prediction to reach the upper boundary of prediction accuracy.
Several considerable issues have been inevitably raised in this study. Most importantly, the training set used in this study may contain a certain amount of incorrect information as some of nonsite residues can be determined as siteresidues in the future. The prediction accuracy for general machine learning (nonparametric) models is highly dependent on the quality of the input/training dataset. If the dataset were to contain inaccurate information, it can significantly affect the learning process thus predicting sites incorrectly. The use of a semiparametric model in this study takes upon assumptions that are stronger than those of nonparametric models so that it can more robustly perform its prediction task on noisedatasets than pure nonparametric methods. However, even when the semiparametric approach was shown to be more resistant to the noisedata, the problems caused by using potentially corrupted dataset cannot be resolved completely unless an experimentally complete and faultless phosphorylation data is provided.
Despite of our effort to collect high quality phosphorylation data for all kind of kinase families, each kinase dataset is excessively biased to a few kinase families. For example, PKA and PKC kinase family data take approximately 69% of AGC kinase group dataset. Hence, a learner trained with AGC kinase group dataset may not perform well with other kinase families in the AGC kinase group as it was trained mostly with PKA or PKC kinase family data. It is ideal if we can collect more experimentally verified data for those kinase families; however, in the case that this is not possible, matched sampling methods were shown to be somewhat effective.
One of the experiments conducted to compare the performance of SiteSeek with other existing phosphorylation site predictors (Table 5) reveals several considerable issues. First, the predictive performance of other predictors can be overestimated in that it is almost infeasible to confirm whether their training datasets contains any of amino acids sequences in our testing set. The presence of testing examples in their training dataset not only violates the assumptions needed for learning but also makes evaluation measures unreliable. Hence, this issue should be taken into account carefully for more accurate evaluation and comparison of models. It should be noted that as the sevenfold cross validation was also used for the evaluation of SiteSeek, no redundant data entry for training and testing sets was found. Second, the window sizes adopted in this study may not be suitable for kinase family or group based predictions. In the literature, it has been discussed that the major determinants for phosphorylation sites are 9 neighbouring residues for tyrosine and threonine; 11 for serine [15]. Several studies using computational methods recently have shown that the optimal window sizes are found to be different depending on types of kinase families. However, no wetlab research has validated the kinase family based optimal window sizes to date. Hence, further research should be carried out so as to provide for a more accurate evaluation and prediction of phosphorylation sites.
Conclusion
This paper identified the effectiveness and utility of our newly proposed machine learning based predictor, SiteSeek for phosphorylation site prediction. This study addressed two important issues in the computational prediction of protein phosphorylation sites. Current encoding schemes like PSSM and Orthogonal Encodings do not provide sufficient information for accurate prediction of phosphorylation sites using existing computational models. Our approach uses compact PSSM with efficient hydrophobicity scale proves to be more effective in the prediction of phosphorylation sites. Next, for a given set of high dimensional protein data, the combination of a parametric local model with a nonparametric global model provided a way of finetuning the model by the adjustment of a single smoothing parameter σ as well as providing efficient semiparametric approximation. This was demonstrated by our above four consecutive experiments. The semiparametric approach used in AdaptiveLEKM was shown to be effective by finding an optimal tradeoff between parametric and nonparametric models with significantly reduced computations. When tested with the newly built PSBenchmark_1 dataset, SiteSeek which uses the AdaptiveLEKM and CEHProfile achieved the best prediction accuracy when compared with contemporary phosphorylation site predictors. Thus, allowing us to accurately predict phosphorylation sites in proteins so as to better understand their functions in biological systems.
Methods
PSBenchmark_1 dataset
The fair comparison and assessment of phosphorylation site predictors is complicated as all use different phosphorylation site datasets in the literature. In this study, we use a newly developed comprehensive dataset, namely PSBenchmark_1 for the purpose of benchmarking sequencebased phosphorylation site prediction methods. It is widely known that accurate classification is highly dependent upon the quality of data sets of both positive and negative examples. However, such a golden standard datasets are not yet available for protein phosphorylation site prediction. PSBenchmark_1 contains experimentally verified phosphorylation sites manually extracted from major protein sequence databases and the literature. The dataset comprises of 1,668 polypeptide chains and as shown in Table 6, the chains are categorised in four major kinase groups, namely cAMPdependent protein kinase/protein kinase G/protein kinase C extended family (AGC), calcium/calmodulindependent kinase (CAMK), cyclindependent kinaselike kinase (CMGC) and tyrosine kinase (TK) groups. The dataset comprises of 513 AGC chains, 151 CAMK chains, 330 CMGC chains, and 216 TK chains. The dataset is nonredundant in a structural sense: each combination of topologies occurs only once per dataset. Protein sequences are taken from the Protein Data Bank (PDB) [31], SwissProt [32], Phospho3D [10], Phospho.ELM [33] and literature.
Compact evolutionary and hydrophobicity profile
One of our preliminary experiments showed that the less discriminative features residing in the original sequence profile can diminish the predictive performance of the learners. So, we utilise the nonlinear autoassociative network to filter the lessdiscriminative features in the dataset. In the process of the filtering, the model finds the optimal data dimension for more accurate prediction. As shown in Figure 2, the reduced data dimension of 17 (dm_17) provides better evaluation measures than its original data dimensions of the sequence profile (dm_20). The best data dimension (dm_17) observed in this experiments divulge that not all the information in the sequence profile is useful for accurate prediction. In other words, some amounts of lessdiscriminative features that can impede recognition of the learner exist in the profile. Hence, it is imperative that to reach the upper boundary of the prediction accuracy, additional informative features from other sources should be added.
The promising additional information can be sought from the widely known hydrophobicity scales. A number of researchers selected hydrophobicity as the main feature among many other physicochemical properties for protein structure prediction (such as polarity, charge or size) [34–36]. More importantly, several recent studies reported that protein hydrophobicity can be affected by the level of phosphorylation or vice versa [37–39]. Hydrophobicity is a very important factor in protein stability. The "hydrophobic effect" is believed to play a fundamental role in the spontaneous folding of proteins. It can be expressed as the free energy (kilocalories per mole) of transfer of amino acid side chains from cyclohexane to water. The amino acids with positive values of free energy in transferring cyclohexane to water are hydrophobic and the ones with negative values are hydrophilic [34]. Table 7 shows hydrophobicity scales, and the hydrophobicity matrix can be formulated using the following function.
Given:
Amino_Acid [] = {C, Q, E, G, H, I, L, K, M, F, P, S, T, W, Y, V,...} and
Hydrophobicity_Index [] = {1.28, 5.54, 6.81, 0.94, 4.66, 4.92, 4.92, 5.55, 2.35, 2.98, 4.04, 3.40, 2.57, 2.33, 0.14, 4.04,...},
where the denominator 20 is used to convert the data range into [0, 1]. Hydrophobicity matrix [3, 4] means the absolute value of the difference of the hydrophobicity indices of two amino acids E (6.81) and G (0.94). With the range adjustment, now we obtain 0.2935.
In the case of structure/function families classification of protein sequences, various hydrophobicity scales were thoroughly examined by David [40]. He showed the effectiveness of numerous hydrophobicity scales, and concluded that the Rose scale [41] was superior to all others when used for protein structure prediction. The Rose scale is correlated to the average area of buried amino acids in globular proteins and is shown in Table 8[41]. However, Korenberg et al. [36], stated several key drawbacks with Rose scale. As it is not a onetoone mapping, different amino acid sequences can have identical hydrophobicity profiles. Also, the scale covers a narrow range of values, while causing some amino acids to be weighted more heavily than others. To overcome these problems, the SARAH1 scale, five bits "state" representation for amino acid was introduced by Korenberg et al.
SARAH1 assigns each amino acid a unique fivebit signed code where exactly two bits are nonzero. SARAH1 ranks twenty possible amino acids according to the Rose hydrophobicity scale (Table 8). Each amino acid is assigned a five bit code in descending order of the binary value of the corresponding code. One of the benefits to using the fivebit code is that the complexity of the classifier can be significantly reduced and can arrange these numbers in thirtytwo possible ways (2^{5} = 32). If the representations with no or all ones, and those with 1 or 4 ones are removed, there are exactly twenty representations left. This leaves just enough representation to code for the twenty amino acids. In the case of window size five, a residue vector has 5*11 = 55 dimensions, which leads to less model complexity than the residue vector using orthogonal encoding (20*11 = 220 dimensions) [27].
The resulting scale in Table 9, where the right half is the negative mirror image of the left half, is referred to as SARAH1. The ten most hydrophobic residues are positive, and the ten least hydrophobic residues are negative. Korenberg et al [36]. indicated that while the above scales carry information about hydrophobicity, scales can similarly be constructed to embed other chemical or physical properties of the amino acids such as polarity, charge, α – helical preference, and residue volume.
In order to create a new profile, we use the idea above in addition to the existing sequence profile generated by PSIBLAST. As suggested in Figure 2, the lessdiscriminatory features in the sequence profile are removed by using the autoassociative network in order to prevent some possible problems that may be caused by the high complexity of the learner. Finally, CEHProfile which contains selected evolutionary information and SARAH1 hydrophobicity scale is created.
Learning in high dimensional space
Protein sequence data can be mathematically viewed as points in a high dimensional space. For example, a sequence of 10 amino acids represents a search space of 20^{10} possibilities and requires a network of 200 inputs. In many applications, the curse of dimensionality is one of the major problems that arise when using nonparametric techniques [42].
Learning in the high dimensional space causes several important problems. First, the good data fitting capacity of the flexible "modelfree" approach often tends to fit the training data very well and thus, have a low bias. However, the potential risk is the overfitting that causes high variance in generalisation. In general, the variance is shown to be a more important factor than the learning bias in poor prediction performance [43]. Second, with the high dimensional data, as the number of hidden nodes of the network is severely increased, it eventually leads to an exponential rise in computational complexity. A high complexity model generally shows a low bias but a high variance [44]. On the other hand, a model with low complexity shows a high bias but a low variance. Hence, a good model balances well between model bias and model variance. This problem is generally regarded as the term "biasvariance tradeoff".
Semiparametric modelling for the biasvariance tradeoff
One of the solutions to the problems above is socalled semiparametric modelling. Semiparametric models take assumptions that are stronger than those of nonparametric models, but are less restrictive than those of parametric model. In particular, they avoid most serious practical disadvantages of nonparametric methods at the price of an increased risk of specification errors.
The proposed model, AdaptiveLEKM takes a form of the semiparametric model and it finds the optimal tradeoff between parametric and nonparametric models. Thus, it can take advantages of both models while effectively avoiding the curse of dimensionality. The AdaptiveLEKM contains the evolutionary information represented within the local model. Its global model works as a collaborative filter that transfers the knowledge amongst the local models in formats of the hyperparameters. Here, as a local model of AdaptiveLEKM, an effective data compression technique is used for the data localisation. The local model contains an efficient vector quantisation method.
Vector quantisation for localityeffectiveness
Vector Quantisation (VQ) is a lossy data compression technique based on the principle of book coding. Its basic idea is to replace with key values from an original multidimensional vector space into values from a discrete subspace of lower dimension. The lowerspace vector requires less storage space and the data is thus compressed.
Consider a training sequence consisting of M source vectors, T = {x_{1}, x_{2},..., x_{ m }}. M is assumed to be sufficiently large and so that all the statistical properties of the source are captured by the training sequence. We assume that the source vectors are kdimensional, X_{ m }= (x_{m,1}, x_{m,2},..., x_{m, k}), m = 1, 2,..., M. These vectors are compressed by choosing the nearest matching vectors and form a codebook consisting the set of all the codevectors. N is the number of codevectors, C = {c_{1}, c_{2},..., c_{ n }} and each codevector is kdimensional, c_{ n }= (c_{n,1}, c_{n,2},..., c_{n, k}), n = 1, 2,..., N. The representative codevector is determined to be the closest in Euclidean distance from the source vector. The Euclidean distance is defined by:
where
x_{ j }= the j^{th}component of the source vector,
c_{ ij }= the j^{th}is components of the codevector c_{ i }.
S_{ n }is the nearestneighbour region associated with codevector c_{ n }, and the partitions of the whole region are denoted by P = {S_{1}, S_{2},..., S_{ N }}. If the source vector X_{ m }is in the region S_{ n }, its approximation can be denoted by Q(X_{ m }) = c_{ n }, if X_{ m }∈ S_{ n }. The Voronoi region is defined by:V_{ i }= {x ∈ R^{k}: x  c_{ i } ≤ x  c_{ j }, for all j ≠ i}
As depicted in Figure 3, the training vectors falling in a particular region are approximated by a red dot associated with that region. To find the most optimal C and P, vector quantisation uses a squareerror distortion measure specifying exactly how close the approximation is. The distortion measure can be given as:
If C and P are solution to the above minimisation problem, then it must satisfy two conditions namely nearest neighbour and centroid conditions. The nearest neighbour condition indicates the subregion S_{ n }should consist of all vectors that are closer to c_{ n }than any of the other codevectors. It is written as:S_{ n }= {x: x  c_{ n }^{2} ≤ x  c_{n'}^{2} ∀ n' = 1, 2,..., N}
The centroid condition requires the codevector c_{ n }should be average of all those training vectors that are in its Voronoi Region S_{ n }.
Kernel machine as an collaborating filter
As a key collaborator of Adaptive LEKM, we use an effective kernel classifier to construct the global model. SVM can be seen as a set of related supervised learning and generalised linear classifiers. The key features of SVMs are the use of kernels, the absence of local minima, the sparseness of the solution and the capacity control obtained by optimising the margin [45]. A significant advantage of SVMs is that whilst ANNs can suffer from multiple local minima, the solution to an SVM is global and unique [45, 46]. SVMs are known as maximum margin classifiers since they classify their objects by minimising the empirical generalisation error and maximising the geometric margin simultaneously. Where the two classes are not separable, they map the input space into a highdimensional feature space (where the classes are linearly separable) by using a nonlinear kernel function. The kernel function calculates the scalar product of the images of two examples in the feature space. Given a ddimensional input vector, x_{ i }= (x_{1}, x_{2},..., x_{ n }) with two labels, y_{ i }∈ {+1, 1} (i = 1, 2,..., N), the hyperplane decision function of binary SVM with kernel method is written as:
and the following quadratic program:
maximise $W(a)={\displaystyle \sum _{i=1}^{\ell}{a}_{i}}\frac{1}{2}{\displaystyle \sum _{i,j=1}^{\ell}{a}_{i}{a}_{j}{y}_{i}{y}_{j}k({x}_{i},{x}_{j})}$
subject to a_{ i }≤ 0, i = 1,..., ℓ, and $\sum _{i=1}^{\ell}{a}_{i}{y}_{i}=0$.
where ℓ is the number of training patters; a_{ i }are the parameters of the SVM; k(·,·) is a suitable kernel function, and b is the bias term.
The SVM used in the AdaptiveLEKM is the modified version of SVM^{light}package. It uses an RBF kernel for the classification and the hyperparameters used in the SVM were optimised using a 7fold crossvalidation (Refer to Section Overall Architecture of SiteSeek). In order to find optimal values for the hyperparameters, a number of values were considered and tested against the newly built PSBenchmark_1 dataset. The optimal values were chosen for the PSBenchamrk_1 dataset were C: 1.5, γ: 0.04, and ε: 0.1.
LocalityEffective Kernel Machine
In the literature, it is claimed that one of the most serious problems with SVMs is the high algorithmic complexity and extensive memory requirements of the required quadratic programming in largescale tasks [47]. As observed in the above equation, SVM extracts worstcase examples x_{ i }and use statistical analysis to build large margin classifiers. However, in AdaptiveLEKM, we use the centroid vector of each voronoi region which can be expressed as:
To construct a semiparametric model, we substitute Q_{ i }(X) for each training sample x_{ i }used in the SVM decision function. The AdaptiveLEKM's approximation can be written as:
and the following quadratic program:
maximise $W(a)={\displaystyle \sum _{i=1}^{\ell}{a}_{i}}\frac{1}{2}{\displaystyle \sum _{i,j=1}^{\ell}{a}_{i}{a}_{j}{y}_{i}{y}_{j}k({Q}_{i}(x),{Q}_{j}(x))}$
subject to a_{ i }≥ 0, i = 1,..., ℓ, and $\sum _{i=1}^{\ell}{a}_{i}{y}_{i}=0$.
The SVM is considered as a purely nonparametric model, whereas the AdaptiveLEKM can be considered as semiparametric model as it adopts the method of grouping of the associated input vectors in each class i. Hence, the performance of proposed model has some advantages in comparison to the pure parametric models and pure nonparametric models in terms of learning bias and generalisation variance especially on high dimensional protein datasets.
As the AdaptiveLEKM uses the centroid vector of each nearest neighbour region, we can obtain the optimal representations by finding right size of each sub regions. In other words, a good tradeoff between parametric and nonparametric can be found by adjusting the size of each subregion. If the feature space is partitioned to too many subregions, the model becomes closer to nonparametric model. So, it is eventually susceptible to overfitting which causes high model variance problem. Contrarily, if the space is divided into too small number of regions, the codevectors cannot correctly represent the original dataset (as they miss too much information). And the model eventually produces high leaning bias. Hence, it is crucial to find a good tradeoff between the parametric and nonparametric models.
In AdaptiveLEKM, the size of Voronoi regions is continually changed until they find the most optimal tradeoff between model bias and generalisation variance. Eventually, the global model is able to transfer the knowledge amongst the optimised local models in formats of the hyperparameters. Our method which includes selection of the corresponding output y_{ i }and the finding optimal size of the associated input regions in each class i is shown to be effective in finding an optimal tradeoff between model bias and variance. However, if we assume that a centroid vector is calculated by its associated cluster and the cluster contains noise vector, the centroid vector may incorrectly represent its associated cluster or regions. Hence, in our method, an autoassociative neuralnetwork [48] is used to perform nonlinear mapping from x to x' to eliminate noise or lessdiscriminatory features that can impede recognition.
Autoassociative network for nonlinear dimensionality reduction
For the nonlinear mapping, we use Scholz's [49] standard autoassociative network as also widely known as nonlinear principal component analysis (NLPCA). The autoassociative network contains three hidden layers between the input and output layers namely encoding, bottleneck and decoding units. In the hidden units as depicted in Figure 4, the input signals are transferred to the "encoding" neurons in the first hidden units. The hyperbolic tangent function is used as the transfer function here, and again when the signal moves from the "bottleneck" neuron in the second hidden layer to the "decoding" neurons in the third hidden layer. Data compression is achieved by the bottleneck, with the single bottleneck neuron giving the leading nonlinear principal component. The numbers of encoding and decoding neurons are adjustable for the optimal fit, however, are set the same for simplicity. The autoassociative network in Figure 4 with 3, 4, 1, 4 and 3 neurons in its 5 layers will be referred to as a 34143 model.
AdaBoost for the network tuning
In order to maximise the performance of the AdaptiveLEKM, we utilised a network boosting method called Adaptive Boosting (AdaBoost). In general, boosting is known as a technique to improve the performance of any base machine learning algorithms. The AdaBoost algorithm was proposed by Freund and Schapire [50] and it was shown to be a solution for many practical difficulties of previous boosting algorithms. Boosting combines weak learners to find a highly accurate classifier or better fit for the training set [51]. In this study, the AdaBoost was modified for the LEKM for the network boosting. As observed in our experiments, the modified AdaBoost was tested with the LEKM and showed that it can fit into its architecture for more accurate prediction of phosphorylation sites. A standard boosting algorithm can be written as:
Given: (x_{1}, y_{1}),...,(x_{ NV }, y_{ NV }) where x_{ i }∈ X, y_{ i }∈ Y = {1, +1}
Initialise D_{1}(i) = 1/NV
For t = 1,...., T:

Find the classifier h _{ t }: X → {1, +1} that minimises the error with respect to the distribution D _{ t }
${h}_{t}=\mathrm{arg}\phantom{\rule{0.1em}{0ex}}\underset{{h}_{j}\in H}{\mathrm{min}\phantom{\rule{0.1em}{0ex}}}{\epsilon}_{j}$ where ${\epsilon}_{j}={\displaystyle \sum _{i=1}^{NV}{D}_{t}(i)[{y}_{i}\ne {h}_{j}({x}_{i})]}$

Get weak hypothesis h _{ t }: X → [0, 1]

Choose a _{ t }∈ R, typically ${a}_{t}=\frac{1}{2}in\frac{1{\epsilon}_{t}}{{\epsilon}_{t}}$ where ε _{ t }is the weighted error rate of classifier h _{ t }

Update:
$${D}_{t+1}(i)=\frac{{D}_{t}(i)\mathrm{exp}\phantom{\rule{0.1em}{0ex}}({a}_{t}{y}_{i}{h}_{t}({x}_{i}))}{{Z}_{t}}$$
where
NV = total number of training vectors,
X = a domain or instance space of each x_{ i }belong to,
Y = a label set of each label y_{ i },
Z_{ t }= a normalisation factor (chosen so that D_{t+1}will be a distribution),
R = its sign is the predicted label {1, +1}.
Output the final hypothesis:
In the training set, each x_{ i }belongs to a domain X, and each label y_{ i }is in a label set Y. Here, the Y should be {1, +1} as phosphorylation sites are indicated as positive (+1) or negative (1) values only. After selecting an optimal classifier h_{ t }for the distribution D_{ t }, the examples x_{ i }that the classifier h_{ t }identified correctly are weighted less and those that it identified incorrectly are weighted more. Therefore, when the algorithm is testing the classifiers on the distribution D_{ t }+1, it will select a classifier that better identifies those examples that the previous classifier missed. At each iteration, the AdaBoost embedded in AdaptiveLEKM constructs weak learners based on this method called weighted examples.
Overall Architecture of SiteSeek
As illustrated in Figure 4, SiteSeek contains four main components. First, given amino acids sequences, SiteSeek utilise PSIBLAST to generate Position Specific Scoring Matrix (PSSM) with an evalue threshold for inclusion of 0.001 and six search iterations of nonredundant (nr) sequence database. The PSSM has 20 × N elements, where N is the length of the target sequence and each element represents the loglikelihood of a particular residue substitution based on a weighted average of BLOSUM62 [52] matrix scores for a given alignment position in the template.
Second, compact PSSM is generated by using the standard autoassociative network. As all the possible less discriminatory features and noises are expected to be eliminated in this step, it is crucial in generating the optimal codebook in the following step. To construct CEHProfile, SARAH1 scales are computed from the amino acid chains in PSBenchmark_1 dataset and are added to the compact PSSM. The CEHProfile which contain compact PSSM and SARAH1 scales was all normalised to fall in the interval [1, 1] by using following algorithm.pn = 2*(pminp)/(maxpminp)  1
where
p = R × Q matrix of input vectors,
minp = R × 1 vector containing minimums for each p,
maxp = R × 1 vector containing maximums for each p.
Third, to find the most optimal parameter C and P as for the solution of the minimisation problem, vector quantisation uses the given distortion measure (See Section Vector Quantisation for LocalityEffectiveness). As discussed in Section LocalityEffective Kernel Machine, the modified vector quantiser in AdaptiveLEKM finds the most suitable number of Voronoi regions by considering model bias and generalisation variance generated in each iteration so that the codebook is generated for the classification.
Fourth, our kernel machine uses the resulting codebook and performs its classification tasks. For the fair comparison of our proposed model, we adopted a seven fold crossvalidation scheme for the model evaluation. The PSBenchmark_1 dataset was divided into seven subsamples. One of the subsamples was used for testing and the rest six were used for training. The testing was conducted seven times for each model. When multiple random training and testing experiments were performed, a model was formed from the six subsamples (training samples). The estimated prediction accuracy is the average of the prediction accuracy over seven different datasets for the models. We used the window size of 9 for tyrosine and threonine, and 11 for serine sites [15]. A window size of 9 means 19 amino acids with the tyrosine, threonine or serine site is located at the centre of the window.
Finally, with the threshold T, the final predictions are simulated from the raw output generated by AdaptiveLEKM. During the postprocessing of the network output, as the network generates the raw outputs which have many local peaks, we modified Liu and Rost's [53] method to filter these raw outputs. First, we determined the threshold for each network outputs according to the length (L) of the protein and to the distribution of raw output values for all residues in that protein. We compiled the 96th percentile of the raw output T_{1} and set the threshold T to
T was set to the threshold that divides phosphorylation site and others. If the value of a residue is above the threshold, the residue is regarded as phosphorylation site. Second, we assigned the central residue as a phosphorylation site if three or more residues were predicted as a phosphorylation site. And all parameters for these filters were developed using the validation set only.
The performance of SiteSeek is measured by the accuracy (Ac: the proportion of truepositive and truenegative residues with respect to the total positives and negatives residues), the sensitivity (Sn: the proportion of correctly predicted phosphorylation site residues with respect to the total positively identified residues), the specificity (Sp: the proportion of incorrectly predicted site residues with respect to the total number of phosphorylation site residues) and correlation coefficient (Cc: It balances positive predictions equally with negative predictions and varies between 1 and 1.). Cc reflects a situation in that a method which predicts every residue to be positive, shows prediction accuracy of 100% in detecting positive sites, however 0% accuracy for negative residues. Hence, high value of Cc means that the model is regarded as a more robust prediction system. In addition to the four measures above, the performance of each model is additionally measured by Type I and Type II Error rates as incorrectly predicted residues can be as valuable as the correctly predicted residues for further modification of the model. Type I Error means experimentally verified unmodified sites that are predicted (incorrectly) to be modified; And Type II Error indicates experimentally verified modified sites that are predicted (incorrectly) to be unmodified. The Sn, Sp, Ac and CC can be expressed in terms of true positive (TP), false negative (FN), true negative (TN) and false positive (FP) predictions.
and
The stepwise procedure we have performed can be summarised as follows:

(1)
Data collection, building a new dataset and preprocessing datasets.

(2)
Profiles construction such as PSSM, Orthogonal encoding, and Sarah1.

(3)
Holdout method was performed to divide the combined dataset into 7 subsets (training and testing sets).

(4)
CEPProfile construction for each subset using the autoassociative network.

i.
The autoassociative network was built with five layers in 34143 model.

ii.
Activation Function Selection: Hyperbolic tangent function.

iii.
The original data dimension is reduced into dm_17 (compact_PSSM).

iv.
Sarah1 scale is added to the compact_PSSM.

(5)
The information obtained in (2) and (3) were combined and normalised to fall in the interval [1, 1] to be fed into networks.

(6)
Assign target levels in each profile.

i.
Positive (1) for phosphorylation site residues and negative (1) for nonsite residues. (For tyrosine or threonine sites, nine neighbouring residues are assigned positive (1) whereas eleven neighbouring resides are positive for serine sites).

(7)
Train each model on training set to create a model

(8)
Simulate each model on test set to obtain predicted outputs.

(9)
Postprocessing was performed to find predicted phosphorylation sites locations.
The procedure from (7) to (9) is performed iteratively until we obtain the most suitable kernel and the optimal hyperparameters for AdaptiveLEKM for the given benchmark dataset.
Conflict of interests
The authors declare that they have no competing interests.
Abbreviations
 AdaptiveLEKM:

Adaptive LocalityEffective Kernel Machine, AGC: cAMPdependent protein kinase/protein kinase G/protein kinase C extended family, ANN: Artificial Neural Network, CAMK: Calcium/calmodulindependent Kinase, CEHProfile: Compact Evolutionary and Hydrophobicity Profile, CDK: Cyclindependent Kinase, CK2: Casein Kinase II, CMGC: Cyclindependent kinaselike Kinase, DT: Decision Tree, GRNN: General Regression Neural Network, KLR: Kernel Logistic Regression, kNN: Kernel Nearest Neighbour, MLP: MultiLayered Perceptron, OE: Orthogonal Encoding, PK: Protein Kinase, PKA: cAMPdependent protein Kinase, PKC: Protein Kinase C, PSIBLAST: Position Specific Iterated Basic Local Alignment Search Tool, PSSM: Position Specific Scoring Matrix, RBFN: Radial Basis Function Neural Network, SARAH1: Simultaneously Axially and Radially Aligned Hydrophobicities, SVM: Support Vector Machine, TK: Tyrosine Kinase.
References
 1.
Cohen P: The origins of protein phosphorylation. Nat Cell Biol 2002, 4(5):E127E130.
 2.
Graves L, Bornfeldt K, Krebs E: Historical perspectives and new insights involving the MAP kinase cascades. Advan Sec Mess Phos Res 1997, 31: 49–62.
 3.
Hunter T: Signaling2000 and beyond. Cell 2000, 10(1):113–127.
 4.
Johnson L, Lowe E, Noble M, Owen D: The eleventh datta lecture. the structural basis for substrate recognition and control by protein kinases. FEBS Letters 1998, 430: 1–11.
 5.
Koliba K, Druker B: Protein tyrosine kinases and cancer. Biochim Biophys Acta 1997, 1333: F217F248.
 6.
Pinna LA, Ruzzene M: How do protein kinases recognize their substrates? Biochim Biophys Acta 1996, 1314: 191–225.
 7.
Xue Y, Zhou F, Zhu M, Ahmed K, Chen G, Yao X: GPS: a comprehensive www server for phosphorylation sites prediction. Nucleic Acids Research 2005, 33: W184W187.
 8.
Blom N, SicheritzPontén T, Gupta R, Gammeltoft S, Brunak S: Prediction of posttranslational glycosylation and phosphorylation of proteins from the amino acid sequence. Proteomics 2004, 4: 1633–1649.
 9.
Hunter T: The Croonian Lecture 1997. The phosphorylation of proteins on tyrosine: its role in cell growth and disease. Phil Trans Roy Soc ser B 1998, 353: 583–605.
 10.
Zanzoni A, Ausiello G, Via A, Gherardini PF, HelmerCitterich M: Phospho3D: a database of threedimensional structures of protein phosphorylation sites. Nucl Acids Res 2007, 35: D229D231.
 11.
Ballif BA, Villen J, Beausoleil SA, Schwartz D, Gygi SP: Phosphoproteomic analysis of the developing mouse brain. Mol Cell Proteomics 2004, 3: 1093–1101.
 12.
Beausoleil SA, Jedrychowski M, Schwartz D, Elias JE, Villen J, Li J, Cohn MA, Cantley LC, Gygi SP: Largescale characterization of HeLa cell nuclear phosphoproteins. Proc Natl Acad Sci 2004, 101: 12130–12135.
 13.
Ficarro SB, McCleland ML, Stukenberg PT, Burke DJ, Ross MM, Shabanowitz J, Hunt DF, White FM: Phosphoproteome analysis by mass spectrometry and its application to Saccharomyces cerevisiae. Nat Biotechnol 2002, 20: 301–305.
 14.
Mann M, Ong SE, Gronborg M, Steen H, Jensen ON, Pandey A: Analysis of protein phosphorylation using mass spectrometry: deciphering the phosphoproteome. Trends Biotechnol 2002, 20(6):261–268.
 15.
Songyang Z, Shoelson SE, Chaudhuri M, Gish G, Pawson T, Haser WG, King F, Roberts T, Ratnofsky S, Lechleider RJ, Neel BG, Birge RB, Fajardo JE, Chou MM, Hanafusa H, et al.: SH2 domains recognize specific phosphopeptide sequences. Cell 1993, 72: 767–778.
 16.
Iakoucheva LM, Radivojac P, Brown CJ, O'Connor TR, Sikes JG, Obradovic Z, Dunker AK: The importance of intrinsic disorder for protein phosphorylation. Nucleic Acids Research 2004, 32(3):1037–1049.
 17.
Obenauer JC, Cantley LC, Yaffe MB: Scansite 2.0: proteomewide prediction of cell signaling interactions using short sequence motifs. Nucleic Acids Research 2003, 31(13):3635–3641.
 18.
Hjerrild M, Stensballe A, Rasmussen TE, Kofoed CB, Blom N, SicheritzPonten T, Larsen MR, Brunak S, Jensen ON, Gammeltoft S: Identification of Phosphorylation Sites in Protein Kinase A Substrates Using Artificial Neural Networks and Mass Spectrometry. Proteome Research 2004, 3: 426–433.
 19.
Kim JH, Lee J, Oh B, Kimm K, Koh I: Prediction of phosphorylation sites using SVMs. Bioinformatics 2004, 20(1):3179–3184.
 20.
Frishman D, Argos P: Incorporation of nonlocal interactions in protein secondary structure prediction from the amino acid sequence. Protein Engineering 1996, 9: 133–142.
 21.
King RD, Sternberg MJE: Identification and application of the concepts important for accurate and reliable protein secondary structure prediction. Protein Science 1996, 5: 2298–2310.
 22.
Rost B, Sander C: Prediction of protein secondary structure at better than 70% accuracy. Journal of Molecular Biology 1993, 232: 584–599.
 23.
Salamov AA, Solovyev VV: Prediction of protein secondary structure by combining nearestneighbor algorithms and multiple sequence alignments. J Mol Biol 1995, 247: 11–15.
 24.
Zvelebil MJ, Barton GJ, Taylor WR, Sternberg MJE: Prediction of protein secondary structure and active sites using alignment of homologous sequence. J Mol Biol 1987, 194: 957–961.
 25.
Yaffe MB, Leparc GG, Lai J, Obata T, Volinia S, Cantley LC: A motifbased profile scanning approach for genomewide prediction of signaling pathways. Nat Biotechnol 2001, 19: 348–353.
 26.
Baldi P, Brunak S: Bioinformaticsthe Machine Learning Approach. MIT Press, Cambridge; 1998.
 27.
Zhang B, Zhihang C, Murphey YL: Protein secondary structure prediction using machine learning. Proc of IEEE Conf on Neural Networks (IJCNN) 2005, 1: 532–537.
 28.
Lin K, May ACW, Taylor WR: Amino Acid Encoding Schemes from Protein Structure Alignments: Multidimensional Vectors to Describe Residue Types. Journal of Theoretical Biology 2002, 216: 361–365.
 29.
Lohmann R, Schneider G, Behrens D, Wrede P: A neural network model for the prediction of membranespanning amino acid sequences. Protein Science 1994, 3: 1597–1601.
 30.
Yang ZR, Chou KC: Biosupport vector machines for computational proteomics. Bioinformatics 2004, 20: 735–741.
 31.
Berman HM, Westbrook J, Feng Z, Gilliland G, Bhat TN, Weissig H, Shindylov IN, Bourne PE: The Protein Data Bank. Nucleic Acids Research 2000, 28: 235–242.
 32.
Amos B: Serendipity in bioinformatics, the tribulations of a Swiss bioinformatician through exciting times! Bioinformatics 2000, 16: 48–64.
 33.
Diella F, Cameron S, Gemund C, Linding R, Via A, Kuster B, SicheritzPonten T, Blom N, Gibson TJ: Phospho.ELM: a database of experimentally verified phosphorylation sites in eukaryotic proteins. BMC Bioinformatics 2005, 5: 79.
 34.
Hu H, Pan Y: Improved Protein Secondary Structure Prediction Using Support Vector Machine with a New Encoding Scheme and an Advanced Tertiary Classifier. IEEE Trans on NanoBioScience 2004, 3(4):265–271.
 35.
Kim H, Park H: Protein secondary structure prediction based on an improved support vector machines approach. Protein Engineering 2003, 16: 553–560.
 36.
Korenberg MJ, David R, Hunter IW, Solomon JE: Automatic classification of protein sequences into structure/function groups via parallel cascade identification: a feasibility study. Ann Biomed Eng 2000, 28(7):803–811.
 37.
Daly NL, Hoffmann R, Otvos L, Craik DJ: Role of Phosphorylation in the Conformation of Peptides Implicated in Alzheimer's Disease. Biochemistry 2000, 39(30):9039–9046.
 38.
Jang HH, Kim SY, Park SK, Jeon HS, Lee YM, Jung JH, Lee SY, Chae HB, Jung YJ, Lee KO, Lim CO, Chung WS, Bahk JD, Yun D, Cho MJ, Lee SY: Phosphorylation and concomitant structural changes in human 2Cys peroxiredoxin isotype I differentially regulate its peroxidase and molecular chaperone functions. FEBS Letters 2006, 580(1):351–355.
 39.
Li C, Ibrahim HR, Sugimoto Y, Hatta H, Aoki T: Improvement of Functional Properties of Egg White Protein through Phosphorylation by DryHeating in the Presence of Pyrophosphate. J Agric Food Chem 2004, 52(18):5752–5758.
 40.
David R: Applications of nonlinear system identification to protein structural prediction. In MSc thesis. MIT, Cambridge; 2000.
 41.
Rose G, Geselowitz A, Lesser G, Lee R, Zehfus M: Hydrophobicity of Amino Acid Residues in Globular Proteins. Science 1985, 229(Suppl):834–838.
 42.
Hardle W, Muller M, Sperlich S, Warwatz A: Nonparametric and Semiparametric models. Springer, New York; 2004.
 43.
Dietterich TG, Bakiri G: Machine Learning bias, statistical bias and statistical variance of decision tree algorithms. In Tech Rep. Dept Comput Sci, Oregon State Univ, Corvallies; 1995.
 44.
Larose DT: Discovering Knowledge in Data. Wiley; 2005.
 45.
ShaweTaylor J, Cristianini N: Kernel Methods for Pattern Analysis. Cambridge University Press; 2004.
 46.
Burgess CJC: A Tutorial on Support Vector Machines for Pattern Recognition. J Data Mining and Knowledge Discovery 1998, 2(2):121–167.
 47.
Horváth G: Neural Networks in Measurement Systems. In Advances in Learning Theory: Methods, Models and Applications. Edited by: Suykens JAK, Horváth G, Basu S, Micchelli C, Vandewalle J. NATOASI Series in Computer and Systems Sciences, IOS Press; 2003:375–402.
 48.
Kramer MA: Nonlinear principal component analysis using autoassociative neural networks. AIChE Journal 1991, 37: 233–243.
 49.
Scholz M: Analysing periodic phenomena by circular PCA. In Proceedings of the Conference on Bioinformatics Research and Development. Volume 4414. Edited by: Hochreiter S, Wagner R. BIRD; 2007:38–47.
 50.
Freund Y, Schapire RE: Experiments with a new boosting algorithm. Machine Learning: Proceedings of the Thirteenth International Conference 1996, 148–156.
 51.
Schapire RE: A brief introduction to boosting. Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence 1999, 1401–1406.
 52.
Radzicka A, Wolfenden R: Comparing the polarities of the amino acids: Sidechain distribution coefficients between the vapor phase, cyclohexane, 1Octanol, and neutral aqueous solution. Biochemistry 1988, 27: 1664–1670.
 53.
Liu J, Rost B: Sequencebased prediction of protein domains. Nucleic Acids Research 2004, 32(12):3522–3530.
 54.
Radzicka A, Wolfenden R: Comparing the polarities of the amino acids: Sidechain distribution coefficients between the vapor phase, cyclohexane, 1Octanol, and neutral aqueous solution. Biochemistry 1988, 27: 1664–1670.
 55.
Rose GD, Geselowitz AR, Lesser GJ, Lee RH, Zehfus MH: Hydrophobicity of amino acid residues in globular proteins. Science 1985, 229: 834–838.
Acknowledgements
This work is partially supported by an Australian Research Council Grant number DP0667266.
Author information
Affiliations
Corresponding author
Additional information
Authors' contributions
PDY developed and implemented the new model (AdaptiveLEKM) and profile (CEHProfile), and drafted the manuscript. YSH prepared datasets, programming in C++, Shell and Perl scripts, and interpreted the results with PDY. BBZ and AYZ edited the manuscript and introduced the problem initially.
Paul D Yoo, Yung Shwen Ho contributed equally to this work.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Yoo, P.D., Ho, Y.S., Zhou, B.B. et al. SiteSeek: Posttranslational modification analysis using adaptive localityeffective kernel methods and new profiles. BMC Bioinformatics 9, 272 (2008). https://doi.org/10.1186/147121059272
Received:
Accepted:
Published:
Keywords
 Phosphorylation Site
 Kinase Family
 Machine Learning Model
 Voronoi Region
 Sequence Profile