 Proceedings
 Open access
 Published:
Improved general regression network for protein domain boundary prediction
BMC Bioinformatics volumeÂ 9, ArticleÂ number:Â S12 (2008)
Abstract
Background
Protein domains present some of the most useful information that can be used to understand protein structure and functions. Recent research on protein domain boundary prediction has been mainly based on widely known machine learning techniques, such as Artificial Neural Networks and Support Vector Machines. In this study, we propose a new machine learning model (IGRN) that can achieve accurate and reliable classification, with significantly reduced computations. The IGRN was trained using a PSSM (Position Specific Scoring Matrix), secondary structure, solvent accessibility information and interdomain linker index to detect possible domain boundaries for a target sequence.
Results
The proposed model achieved average prediction accuracy of 67% on the Benchmark_2 dataset for domain boundary identification in multidomains proteins and showed superior predictive performance and generalisation ability among the most widely used neural network models. With the CASP7 benchmark dataset, it also demonstrated comparable performance to existing domain boundary predictors such as DOMpro, DomPred, DomSSEA, DomCut and DomainDiscovery with 70.10% prediction accuracy.
Conclusion
The performance of proposed model has been compared favourably to the performance of other existing machine learning based methods as well as widely known domain boundary predictors on two benchmark datasets and excels in the identification of domain boundaries in terms of model bias, generalisation and computational requirements.
Background
Since proteins provide some of the most fundamental information about many processes in almost all organisms, the ability to predict protein structure and function has become one of the most important goals in bioinformatics research [1, 2]. Protein domains represent one of the most useful avenues for the understanding of protein function and domain familybased analysis, and are of great importance in the study of individual proteins [3]. They have been described as units of compact structure [4], function and evolution [5] and folding [6]. In nature, several domains can be formed together with a vast number of possibilities leading to multidomain and multifunctional proteins [7]. Each domain in a multidomain protein has its own functions and jointly works with its neighbours. They serve as modules to form large assemblies like viral particles or provide specific catalytic and binding sites such as those found in enzymes or regulatory proteins.
Over the last three decade, a large number of methods using the three dimensional coordinates of the protein structure have been proposed for more accurate delineation of domains boundaries [8]. Recently, several methods developed to identify domains in globular proteins from one dimensional atomic coordinates and have been gaining much attention. These methods are based on the assumption that a domain has relatively more contacts within itself than with residues in the remainder of the structure [9]. The most recent sequence based methods have been built on the basis of various machine learners.
DOMpro [10] uses evolutionary information (geneexon shuffling), secondary structure and solvent accessibility information with a recursive neural network; CHOPnet [3] utilises evolutionary information, amino acid composition and amino acid flexibility analysed with a neural network; SnapDRAGON [11] predicts domain by using an ab initio protein folding method; DomSSEA [12] uses predicted secondary structure; the Nagarajan and Yona's [13] method is based on analysing multiple sequence alignments from database searches, position specific physiochemical properties of amino acids and predicted secondary structures analysed with a neural network; Galzitskaya and Melnik [9] use side chain entropy of a region to predict domain boundaries; SSEPDomain [14] predicts domains by combining information of secondary structure element alignments, profileprofile alignments and pattern searches; Armidillo [15] uses the amino acids composition to predict domain boundaries; DomCut [16] uses a linker index deduced from a set of domain/linker segments; PRODO [17] uses a neural network method and finally DomainDiscovery [18] uses support vector machines from sequence information including domain linker index. Many of these methods focus exclusively on predicting boundaries for twodomain chains. The overall accuracy of sequencebased methods has been reported in the range of 50 to 70% for single domain proteins and considerably less (<40%) when limited to multidomain proteins [19, 20].
Although a large number of machine learning based methods have showed their superior performance in domain boundary prediction, several important issues that can degrade the performance of machine learning or statisticalbased methods have largely been ignored. It has been widely recognised that high dimensionality of protein sequence data not only causes a dynamic increase in computational complexity but also can be induced into the generalisation problem of nonparametric methods. With machine learning models, better generalisation and faster training (computationally efficient) can be achieved when they have fewer weights to be adjusted by fewer inputs.
This study aims to develop an accurate and reliable protein domain boundary prediction model using stateoftheart machine learning techniques. In this paper, we propose a new semiparametric method where a linear dimensionality reduction method and a nonparametric model are integrated, and show great performance in terms of prediction bias, variance and computational requirements. In the prediction, we use inter domain linker regions with positionspecific scoring matrix (PSSM) generated from PSIBLAST [21], secondary structural information and relative solvent accessibility data. Domain linkers have shown to be useful in cooperative interdomain interactions [22] function regulation [23], protein stability [24], folding rates [25] and domaindomain orientation [26]. The others were adopted based on the assumption that secondary structure elements and the level of solvent accessibility in the boundary regions are different from those found in the rest of the protein. The novel features of this research are the use of a new semiparametric model and importantly, a unique training set (Benchmark_2) built on the consensus of various experts in protein structure. In the literature, as the word "protein domain" is used in various contexts, we adopt the CASP7 definition of protein domain in this study. Domains in CASP7 should fall somewhere in the middle of following three definitions. (1) Conserved structural entities with a hydrophobic core. (2) Regions that share a common fold, have some functional similarity, and may be evolutionarily related. (3) A subunit of a polypeptide chain that can fold into a stable tertiary structure independently of any other domain. However, predictors are only able to evaluate predictions made for the first two definitions.
Results
A fair and precise comparison of domain boundary predictors is complicated by the fact that the existence of various domain datasets/databases which may conflict and are biased in some cases [27]. For a fair comparison of the proposed model to other current methods, we have tested each model with Benchmark_2 dataset, a new comprehensive dataset that was developed for the purpose of benchmarking structurebased domain identification methods [28] and CASP7 dataset, the most widely known benchmark in protein structure prediction.
The objectives of this study are twofold. The first is to show the effectiveness of the proposed model by accurately comparing it with stateoftheart machine learning models in terms of prediction ability, stableness/robustness, and computational efficiency. At the same time, we find the most suitable window size for the given datasets. Second, by using the best window size, we compare the predictive performance of the proposed model with existing and widely known domain boundary predictors on CASP7 benchmark dataset.
With the first objective, we use the Benchmark_2 dataset and adopt a seven fold crossvalidation scheme for our model evaluation. We evaluate the performance of each model by looking at the average prediction accuracy, the variance of the accuracy, the average training and testing time obtained over our four different datasets constructed based on different window sizes (win_7, win_11, win_19 and win_27). The performance of our proposed model are compared with other general Artificial Neural Network (ANN) models such as MultiLayered Perceptron (MLP), Radial Basis Function Network (RBFN) and General Regression Neural Network (GRNN).
In order to validate the effectiveness of our proposed model, the model should be able to achieve a good tradeoff between the model bias and generalisation variance (Figure 1). Table 1 presents the average prediction accuracy, the variance of prediction accuracy of each prediction model on the seven testing samples for four different window sizes as well as the average training and testing time of each model. As observed, IGRN produced the comparable predictive performance, the overall prediction accuracy of 64.79% and the smallest variance (0.86) than that of GRNN, RBFN and MLP. In the experiment results above, the semiparametric approach of IGRN was demonstrated to be effective for stable prediction. Although our semiparametric approach (IGRN), the incorporation of reduced dimension of input vectors into pure nonparametric model has brought better generalisation and less computational requirements, the prediction accuracy of the proposed model is slightly less accurate (64.79%) than original model, GRNN (65.36%) on the Benchmark_2 dataset.
In order to reduce the learning bias of the IGRN, we adopted a boosting algorithm into the IGRN and it eventually showed much better tradeoff between learning bias and generalisation variance that that of IGRN. IGRN II (AdaBoost embedded in IGRN) achieved the best prediction accuracy as well as comparably low variance among different neural network models. As for the computational efficiency, IGRN and IGRN II both showed comparable performance over other three machine learners. It means our semiparametric approach of proposed model not only has more stable prediction capability than other ANN models but also demands less computation requirements.
In Table 1, we can observe that the highest accuracy was found in the window size of seven for all models and gradually became less as the size of window got longer. This is in some way an expected result as amino acid residues residing closer to the linker residue contain more structural information which involves stronger shortrange interaction signals than others.
With the result obtained from above experiment, the predictive performance of the proposed model is now compared with the most widely known domain boundary predictors, namely, DOMpro, DomPred, DomSSEA, DomCut and DomainDiscovery. DOMpro [10] uses machine learning algorithms, in the form of recursive neural networks and predicts protein domains using a combination of evolutionary information in the form of profiles, predicted secondary structure, and predicted relative solvent accessibility. DomPred uses a combined homology and fold recognitionbased approach. The sequence homology approach simply attempts to distinguish domain boundaries from overlapping edges in PSIBLAST multiple sequence alignments. The fold recognition approach relies on secondary structure element alignments, using the DomSSEA method [12], in order to find domain boundaries in more distant homologs. The method has an accuracy of 49% at predicting the domain boundary location within 20 residues using a representative set of two domain chains. DomCut [16] predicts linker regions based on sequence alone, relies solely on amino acid propensity. This method simply defines a linker region to be one that has lower linker index values than a specified threshold value. The linker index of amino acids are computed from a database of about 235 nonredundant (pairwise homology <30%) protein sequences. Only the continuous domain (functional or structural) segments of size 50 to 500 are chosen provided the linker region between them is of the size 10 to 100 residues. Linker index of an amino acid is the log ratio of its frequency of being present in a linker region and the frequency presence in a domain region. As used in DomCut, IGRN will also employ domain linker index to model linker regions. Finally, DomainDiscovery [2] is newly developed machine learning based domain predictor. It uses Support Vector Machines (SVMs) and additional sequencebased information, domain linker index during its training. It has shown its comparable predictive performance to other widely known domain predictors in the prediction of domain boundary.
Table 2 shows the experiment results of five domain boundary predictors and our proposed models on CASP7 dataset. The dataset contains 64 1domain chains, 21 2domain chains and 2 3domain chains. IGRN models have shown the comparable performance to other domain boundary predictors as well as its original model (GRNN). IGRN II showed the best prediction accuracy amongst the stateoftheart predictors by reaching the prediction accuracy of 69.44%. IGRN II correctly predicted at 87.93%, 19.12%, and 7.50% accuracy for 1domain, 2domain, and 3domain chains respectively. Interestingly, although the overall accuracy of IGRN (68.10%) is slightly less than that of GRNN (68.52%), its performance for 2domain and 3domain was about 7 points higher than that of GRNN. It may indicate that the semiparametric approach of IGRN which brings more stable prediction seems more suitable for domain boundary prediction problem. In addition, IGRN II also showed superior performance on multidomain chains. The stability/generalisation ability of our proposed models was again proved with the CASP7 dataset. However, it is noted that the prediction results for the 3domain chains are not statistically significant as the CASP7 dataset contains only two 3domain protein chains. The prediction performance of the proposed model (IGRN) should be further tested on more 3domain or larger protein chains.
Figure 2 shows that the comparison of prediction scores simulated by IGRN and GRNN on a protein chain, CASP7 target number: T0318 (PDB code: 2HB6). The protein chain has two domains and its boundary is at the residue 155. IGRN gives clear indication of the domain boundary at the residue 155 and its signal is far stronger than that of GRNN by reaching almost 1 point (0.9964). However, GRNN's signal at the domain boundary is generally only around at 0.3 point which may be indistinguishable from neighbouring signals. Evidently, IGRN offers an additional level of advantages over GRNN and other existing predictors by providing more clear and strong indication of boundary location.
Discussion
Although various machine learning based domain predictors have been developed, their limited stability/generalisation ability has revealed their inadequacy for domain boundary prediction. Our proposed semiparametric approach was seemed to provide a good tradeoff with both low bias and low variance. The experimental results confirmed our hypothesis that the IGRN produces improved generalisation while it is preserving the accurate data modelling in domain boundary prediction. However, as the prediction accuracy of our proposed model was slightly less (64.79%, IGRN) than that of original model (65.36%, GRNN) on two benchmark datasets, this issue should be taken into account.
From our extensive analysis, it may be caused by the linear nature of the embedded dimensionality reduction method, Principal Components Analysis (PCA). Most of the classical methods perform a linear transformation of the original feature vectors. PCA is the most widely known linear feature extraction techniques. However, it has been shown that its effectiveness is limited by the fact that it is globally linear methods. In other words, the data sets may contain essential nonlinear structures that are invisible to PCA. It is generally believed that protein sequence data contains nonlinear relationships which may be generated from longrange interactions. Thus, a new IGRN incorporated with an efficient nonlinear dimensionality reduction method should be tested in the prediction of domain boundary.
One issue with current protein encoding methods used as input to the predictors is that they may provide insufficient structural information involving various interactions (short, middle and longrange interactions of protein). The formation of protein secondary structure, especially the regions of Î²sheets, involves longrange interactions between amino acids [29]. With current methods, the prediction accuracy for Î²strands is less than that for Î±helix or coil. The Î²sheet formation is seen as a tertiary structure interaction which brings two or more strands together by hydrogen bonds. However, they can be situated far apart in the amino acids sequence. In this case, the residues that are close in three dimensional space occupy distant positions in one dimensional sequence [30]. For statistical or machine learners, it is therefore crucial that the model should be further modified or tuned up to efficiently tackle the problem of longrange interaction or an efficient sequence encoding method that effectively represents various interactions should be developed for more accurate prediction of protein domain boundary.
Another considerable issue is that in general the prediction accuracy of sequencebased methods has been far less (<40%) for multidomain proteins. For example, Liu and Rost's [27] experiments on CATH and SCOP assigned domains to random subsets of 1187 proteins of known highresolution structure and less than 10% sequence homology, showed correct prediction of the number of domains (single or multi) in 69% cases. This accuracy for multidomain case alone was however, only 38%. For the twocontinuous domains proteins, averageaccuracy of dbp prediction in different validation runs is 46% to 51% considering a prediction to be correct if it were in Â±20 residues interval of the CATH and SCOP assigned boundaries. Our experiments on CASP7 dataset showed corresponding results as above.
Joshi [20] discussed the main reasons for the problems in deciphering the multidomain protein structures and its possible solutions. With experimental data, although the structure within a domain is fixed, the relative positioning of two domains within the same chain can vary. For this reason, and for the fact that protein structural domains are independent folding units, it is unusual to find single crystal structures containing more than one domain. For similar reasons, protein modelling by database searching, sequence alignment and/or phylogenic analysis and etc is also better performed on a single domain, rather than a multidomain polypeptide. Hence, in most cases, the number of domain in a protein should be firstly identified to determine the locations of such domains on the primary chain before embarking on any standard method of protein structure/function determination. The identification of linker regions connecting two distinct domains is also useful in finding domain boundary locations accordingly, several domain boundary predictors such as DomCut and DomainDiscovery employing domain linker information showed reasonably better predictive performance in domain boundary prediction.
Conclusion
This paper identified the effectiveness and utility of the newly proposed machine learning model, namely Improved General Regression Network (IGRN) for protein domain boundary prediction. For a given set of high dimensional protein data, the combination of a linear principal component local model with a nonparametric global model provided a way of finetuning the model by the adjustment of a single smoothing parameter Ïƒ as well as providing efficient semiparametric approximation. This was demonstrated by the two consecutive experiments. The semiparametric approach used in IGRN was shown to be effective by finding an optimal tradeoff between parametric and nonparametric models with less computational requirements. Finally, with the CASP7 benchmark dataset, IGRN II achieved the best prediction accuracy amongst the contemporary domain boundary predictors by showing its usefulness especially in multidomain chains. However, as the prediction accuracy of IGRN is still limited to less than 70% for multidomain protein datasets, there are still large rooms for further improvement. As discussed, the mid and longterm dependencies of amino acids play key roles for more accurate delineation of domain boundaries. Hence, further research will be carried out on the development of an efficient algorithm to rapidly sift through huge amounts of protein data with the strong capability to capture the mid and longterm information that reside in amino acids.
Methods
Datasets
In this study, we use two benchmark datasets, namely Benchmark_2 and CASP7. Benchmark_2 is a newly developed comprehensive dataset for the purpose of benchmarking structurebased domain identification methods. The Benchmark_2 is similar to the dataset published by Holland et al. [28]. It contains proteins of known structure for which three methods (CATH [31], SCOP [32] and literature) agree on the assignment of the number of domains. The dataset comprises of 315 polypeptide chains, 106 1domain chains, 140 2domain chains, 54 3domain chains, 8 4domain chains, 5 5domain chains and 2 6domain chains. The dataset is nonredundant in a structural sense: each combination of topologies occurs only once per dataset. Sequences of protein chains are taken from the Protein Data Bank (PDB) [33]. Secondary structure information and solvent accessibility are predicted for each chain in the Benchmark_2 using SSpro [34] and ACCpro [35]. Evolutionary information for each chain is obtained by Position Specific Scoring Matrix (PSSM), which was constructed using PSIBLAST [21]. The interdomain linker index was taken from DomCut [16].
CASP7 is one of the most widely known benchmark dataset in domain boundary prediction. Annually, most of the well known domain predictors participate in the Critical Assessment of Techniques for Protein Structure Prediction (CASP) competition. Further information on the available datasets can be found at http://predictioncenter.org/casp7/. The dataset contains 66 1domain chains, 26 2domain chains and 2 3domain chains. For each chain we obtained secondary structure information, solvent accessibility, PSSM and Interdomain linker index by using the previously mentioned methods.
The two raw datasets, Benchmark_2 and CASP7 which contain secondary structure information, solvent accessibility, PSSM and interdomain linker index were all normalised to fall in the interval [1, 1] by using following algorithm.pn = 2*(pminp)/(maxpminp)  1
where
p = R Ã— Q matrix of input vectors,
minp = R Ã— 1 vector containing minimums for each p,
maxp = R Ã— 1 vector containing maximums for each p.
Proposed model
Protein sequence data can be mathematically viewed as points in a high dimensional space. For example, a sequence of 10 amino acids represents a search space of 20^{10} possibilities and requires a network of 200 inputs [36]. Learning in the high dimensional space raises several important problems. For example, the large network training requires a large dataset of known examples, which leads to an exponential rise in computational complexity and susceptibility to the overfitting problem (models obtained from high dimensional data fit the training data very well but perform poorly on previously unseen data).
A good dimensionality method eliminates noise or lessdiscriminatory features that can impede recognition, leaving only a subset of the original features which retain sufficient information to discriminate well among classes. Thus, the number of features used in classification can be reduced when maintaining acceptable classification accuracy. The performance of the most nonparametric statistical classifiers can be extremely dependent upon their input data.
With ANNs, better generalisation and faster training (computationally efficient) can be achieved when they have fewer weights to adjust by fewer inputs. In other words, beyond a certain point, adding new features can actually lead to a reduction in the performance of the classification system [37]. Hence, with a suitable amount of the relevant features, classifiers can not only improve classification speed and efficiency, but also achieve higher accuracy reducing estimation errors associated with finite sample size effects [38].
One solution to the problems above can be semiparametric modelling. Semiparametric models take assumptions that are stronger than those of nonparametric models, but, are less restrictive than those of parametric model. In particular, they avoid most serious practical disadvantages of nonparametric methods but at the price of an increased risk of specification errors. The proposed model, Improved General Regression Network (IGRN) takes a form of the semiparametric model and it finds the optimal tradeoff between parametric and nonparametric models. So, it can have advantages of both models while effectively avoiding the curse of dimensionality.
The IGRN contains the evolutionary information compactly represented within the local model. Its global model also works as a collaborative filter that transfers the knowledge amongst local models in formats of hyperparameters. The local model contains widely known linear principal components. As a collaborative filter, we use General Regression Neural Network. In the literature, GRNN was shown to be effective in noisy environments as it deals with sparse data effectively. Generally, it provides more accurate predictive performance than other existing neural network models. A standard version of the GRNN equation is as follows:
where
x_{ i }= training vector for class i in the input space,
Ïƒ = single learning or smoothing parameter chosen during network training,
y_{ i }= scalar output related to {\underset{\xc2\xaf}{x}}_{i}, and
NV = total number of training vectors.
Figure 3 shows the most general GRNN architecture where f_{ i }(x) represents an arbitrary radial basis function. The above equation is represented using a Gaussian radial basis function f_{ i }(x) as defined by the following equation.
There are many other radial basis functions that can be used instead of the Gaussian function; however, the radial basis function is commonly chosen due to its computational consideration.
As GRNN is known as a universal approximator for smooth function, it should have the capability to solve any smooth function approximation problems when enough data are provided. However, one drawback is that like general kernel methods, it suffers from the curse of dimensionality. It is unable to ignore irrelevant inputs without major modifications to the basic algorithm. Another major issue with GRNN is its large computational requirements. In the above equation, it incorporates every training vector pair {x_{ i }â†’ y_{ i }} into its architecture, ({\underset{\xc2\xaf}{x}}_{i} is a single training vector in the input space, and y_{ i }is the associated desired scalar output). This can require a very large memory space and computations. The two above mentioned issues can significantly degrade the performance of learner when it is applied to high dimensional protein data.
Our method which embeds an efficient linear dimensionality reduction method into the GRNN leads to a more effective approach. If we assume that for each local region of the input space can be represented by reduced vectors generated by an efficient dimensionality reduction method and there is a corresponding scalar output y_{ i }into which it maps; then the GRNN can be approximated within acceptable accuracy. This approach not only achieves the network reduction by efficiently reducing the dimensions of input vectors but also creates a more compact network. This method is theoretically proven to be effective for high dimensionality problems dealing with protein data. The general algorithm for IGRN is:
where
x = input vector,
p_{ i }= transformed input vector for class i in the input space,
Ïƒ = single learning or smoothing parameter chosen during network training
y_{ i }= scalar output related to x_{ i },
Z_{ i }= no. of vectors x_{ i }associated with each p_{ i },
NI = number of unique transformed input vectors p_{ i }.
GRNN is considered as an overfitting and purely nonparametric model, whereas IGRN can be considered as semiparametric model as it contains the evolutionary information represented with the local model. Hence, the performance of proposed model has some advantages in comparison to the pure parametric models and pure nonparametric models in terms of learning bias and generalisation variance especially on high dimensional protein datasets.
In order to maximise the performance of IGRN, we utilise a network boosting method called Adaptive Boosting (AdaBoost). In general, boosting is known as a technique to improve the performance of any base machine learning algorithms. The AdaBoost algorithm was proposed by Freund and Schapire [39] and it was shown to be a solution for many practical difficulties of previous boosting algorithms. Boosting combines weak learners to find a highly accurate classifier or better fit for the training set [40]. In this study, the AdaBoost was modified for the IGRN for the network boosting. As observed in our experiments, the modified AdaBoost was tested with the IGRN and showed that it could fit into its architecture for the more accurate delineation of domain boundaries. A standard boosting algorithm can be written as:
Given: (x_{1}, y_{1}),...,(x_{ NV }, y_{ NV }) where x_{ i }âˆˆ X, y_{ i }âˆˆ Y = {1, +1}
Initialise D_{1}(i) = 1/NV
For t = 1,...,T:

Train weak learner using distribution D_{ t }

Get weak hypothesis h_{ t }: X â†’ R

Choose a_{ t }âˆˆ R

Update:
{D}_{t+1}(i)=\frac{{D}_{t}(i)\mathrm{exp}(\xe2\u02c6\u2019{a}_{t}{y}_{i}{h}_{t}({x}_{i}))}{{Z}_{t}}
where
NV = total number of training vectors,
X = a domain or instance space of each x_{ i }belong to,
Y = a label set of each label y_{ i }
Z_{ t }= a normalisation factor (chosen so that D_{t+1}will be a distribution),
R = its sign is the predicted label {1, +1}.
Output the final hypothesis:
The boosting algorithm takes the training set (x_{1}, y_{1}),...,(x_{ NV }, y_{ NV }) where each x_{ i }belongs to a domain X, and each label y_{ i }is in a label set Y. Here, the Y should be {1, +1} as domain boundaries are indicated as positive (+1) or negative (1) values only. AdaBoost calls the base learning algorithm (IGRN) repeatedly in a series of rounds t = 1,...,T. And it maintains a distribution or set of weights over the training set. The weight of this distribution in training example i in round t is denoted D_{ t }(i). All weights are set equally in the initial state, but on each round, the weights of incorrectly classified examples are increased so that the learner is forced to focus on the hard examples in the training set. The weak learner's job is to find a weak hypothesis h_{ t }: X â†’ {1, +1} appropriate for the distribution D_{ t }. The goodness of a weak hypothesis is measured by its error:
In above equation, the error is measured with respect to distribution D_{ t }on which the learner was trained. In some cases where a learner cannot use the weights D_{ t }in the training examples, a subset of the training examples can be sampled according to D_{ t }, and these unweighted examples can be used to train the learner. In order to validate the effectiveness of the boosting algorithm in the experiments, we IGRN II will indicate AdaBoost embedded in IGRN.
Other machine learning models
In this study, we also compared the performance of the proposed model with other popular neural network models namely MultiLayered Perceptron (MLP) and Radial Basis Neural Network (RBFN) and General Regression Neural Network (GRNN). Here, we look at some basic theory of these models and discuss some general issues.
Feed Forward MultiLayered Perceptron is the most prominent network architecture and primarily used for general classification and regression, implementing feedforward, supervised and heteroassociative paradigm. In MLP, all nodes and layers are arranged in a feedforward manner and its signals travels from input layer to output layer only while FBNN allow signals travelling in both directions [41]. The most general feedforward equation of a three layer MultiLayered Perceptron (MLP) is as follows:y = f_{2}(w_{2}f_{1}(w_{1}x))
where
x = (x_{1}, x_{ x },...,x_{3}), a vector of n predictive or attribute variables,
y = output vector,
w_{1} = matrix of linking weights from input to hidden layer,
w_{2} = matrix of liking weight from hidden to output layer,
f_{1} = transfer function for hidden node,
f_{2} = transfer function for output node.
The most popular choice for f_{1} and f_{2} is the sigmoid function:f_{1}(x) = f_{2}(x) = (1 + e^{x})^{1}
The aim of training of network is to estimate the weight metrics in the above general MLP equation such that an overall error measure such as the mean squared errors (MSE) or sum of squared errors (SSE) is minimised. MSE can be defined as
where
a_{ j }= target value for j^{th}training pattern,
y_{ j }= network output for j^{th}training pattern,
N = number of training patterns.
MLPs had long been a dominating method in various bioinformatics applications such as prediction of secondary structure, tertiary prediction, domain boundary prediction and so forth.
Like MLP, RBFN is also multilayer feedforward network; however, it contains only one hidden layer and the signal in this network is passed forward only. In recent years, RBFN has been widely used due to its structural simplicity and training efficiency.
RBFN combines two different types of learning: supervised and unsupervised. First, at the hidden layer training, one conjoins the input vector into several clusters (unsupervised learning). Afterward, at the output layer training, the output of the RBF network is determined by supervised learning. While for the supervised learning, we have both independent variables and response variables, the unsupervised learning must work without the knowledge of response variables.
RBFN has radial basis functions in hidden nodes and linear functions in output nodes, with adjustable weights that exist only between the hidden and output layers. The output of the j^{th}hidden neuron can be written as:
where
h_{ j }= output of the j^{th}neuron,
Ï• = nonlinear radial basis function,
X = input vector,
C_{ j }= centre of the neuron,
Ïƒ_{ j }= centre spread parameter.
RBFN builds a local approximation by using radial basis functions that are exponentially decaying nonlinear functions while Feed Forward MLP show a global approximation of nonlinear mapping. In general, RBFN was shown to be a good approach for interpolating scattered data and have been applied in various fields. In RBFN, all input spaces are expressed by the overlapping of receptive fields. It consists of three layers, referred to as input, hidden and output layers. The hidden layer contains several Radial Basis Functions (RBFs) and the Gaussian basis function is the most frequently used RBFs [42].
where
x  c_{ j } = Euclidean distance between an input vector x and a centre c_{ j },
Ïƒ_{ j }= the spread of the j^{th}basis function.
In addition to the Gaussian basis functions, other functional forms can be used for the radial basis functions including some types of thinplate spline functions, multiquadratic functions and sigmoidal functions [42]. Any function can be represented in two ways, either as the weighted sum or weighted average. The weighted sum can be presented as:
where
y_{ k }(x) = the k^{th}output,
w_{ jk }= weights between the hidden and output layers,
h_{ j }(x) = output of the j^{th}hidden unit.
The weighted average can be presented as:
It has a higher degree of computational complexity; however, shows better interpolation properties than the weighted sum [43]. The sum of squared error (SSE) between the true values and approximated values from a function is:
which is the sum of residuals of each observation (or sample).
RBFNs usually have three kinds of parameters: the centre and weight of each radial basis function and the weights between the hidden and output layers. Thus, fast and precise identification of these parameters is of primary concern. The centre and width are nonlinear parameters, whereas the weight is linear. Generally, two learning strategies have been applied to RBFNs, nonlinear training methods and hybrid learning methods.
Training, testing and validation
In order to thoroughly test the performance of our proposed model, we conducted two extensive experiments on two benchmark datasets. For the experiment on the Bechmark_2 dataset, we adopted a crossvalidation scheme for the model evaluation. As with jittering, crossvalidation is regarded as one of the effective approaches to improve generalisation of a model. We adopted seven fold cross validation method and random data set selection and testing was conducted seven times for each window size dataset. When multiple random training and testing experiments were performed, a model was formed from each training sample. The estimated prediction accuracy is the average of the prediction accuracy for the models and each window size, derived from the independently and randomly generated test divisions. If the predicted domain boundary is in the range of Â±15 residues of the true domain boundary, we consider it as a correct prediction [18]. We tested four different window sizes (7, 11, 19 and 27) for each model and window size 7 showed the best accuracy for all predictors.
During the postprocessing of the network output, as the network generates the raw outputs which have many local peaks, we filtered these raw outputs by using Liu and Rost's [27] method. We determined the threshold for each network outputs according to the length (L) of the protein and to the distribution of raw output values for all residues in that protein. We compiled the 92nd percentile of the raw output T1 and set the threshold T to
T was set to the threshold that divides domain boundaries and others. If the value of a residue is above the threshold, the residue is regarded as domain boundary. Second, we assigned the central residue as a domain boundary if three or more residues were predicted as a domain boundary. And all parameters for these filters were developed using the validation set only.
As for the experiment on CASP7 dataset, we trained each network using the window size 7 training sets obtained from the experiment on the Benchmark_2 dataset. Each protein chain in CASP7 dataset was fed into each network and a prediction was made. Also identical postprocessing procedure was conducted as performed in the experiment on the Benchmark_2 dataset. With the CASP7 dataset, if the predicted domain boundary is in the range of Â±15 residues of the true domain boundary (CATH assigned dbp) then we consider it as a correct prediction.
The stepwise procedure we have performed can be summarised as follows:

(1)
Data collection, building a new dataset and preprocessing datasets.

(2)
Obtain protein structural information such as PSSM, solvent accessibility, secondary structure and linker index by utilising PSIBLAST [21], ACCpro [34], SSpro [35] and DomCut [16].

(3)
The information obtained in (2) was combined and normalised to fall in the interval [1, 1] to be fed into networks.

(4)
Assign target levels: positive (1) for boundary residues and negative (1) for nonboundary residues.

(5)
Holdout method was performed to divide the combined dataset into 7 subsets (training and testing sets).

(6)
Train each model on training set to create a model.

(7)
Simulate each model on test set to obtain predicted outputs.

(8)
Postprocessing was performed to find predicted boundary locations.
Abbreviations
 ANN:

Artificial Neural Network
 GRNN:

General Regression Neural Network
 IGRN:

Improved General Regression Network
 MLP:

MultiLayered Perceptron
 PSI:

BLAST Position Specific Iterated Basic Local Alignment Search Tool
 PSSM:

Position Specific Scoring Matrix
 RBFN:

Radial Basis Function Neural Network
References
Baldi P, Brunak S: Bioinformaticsthe Machine Learning Approach. MIT Press, Cambridge; 1998.
Sikder AR, Zomaya AY: An Overview of Protein Folding Techniques: Issues and Perspectives. International Journal of Bioinformatics Research and Application 2005, 1(1):121â€“143.
Copley RR, Doerksa T, Letunica I, Borka P: Protein domain analysis in the era of complete genomes. FEBS Letters 2002, 513: 129â€“134.
Richardson JS: The anatomy and taxonomy of protein structure. Adv Protein CHem 1981, 34: 167â€“339.
Bork P: Shuffled domains in extra cellular proteins. FEBS Lett 1991, 286: 47â€“54.
Wetlaufer DB: Nucleation, rapid folding, and globular intrachain regions in proteins. Proc Natl Acad Sci USA 1973, 70: 697â€“701.
Chothia C: Proteins. One thousand families for the molecular biologist. Nature 1992, 357: 543â€“4.
Veretnik S, Shindyalov IN: Computational Methods for Domain Partitioning in Protein Structures. In Computational Methods for Protein Structure and Modeling. Edited by: Xu Y, Xu D, Liang J. SpringerVerlag; 2006.
Galzitskaya OV, Melnik BS: Prediction of protein domain boundaries from sequence alone. Protein Science 2003, 12: 696â€“701.
Cheng J, Sweredoski M, Baldi P: DOMpro: Protein Domain Prediction Using Profiles, Secondary Structure, Relative Solvent Accessibility, and Recursive Neural Networks. Data Mining and Knowledge Discovery 2006, 13(1):1â€“10.
George RA, Heringa J: SnapDRAGON: a Method to Delineate Protein Structural Domains from Sequence Data. J Mol Biol 2002, 316: 839â€“851.
Marsden RL, McGuffin LJ, Jones DT: Rapid protein domain assignment from amino acid sequence using predicted secondary structure. Protein Science 2002, 11: 2814â€“2824.
Nagarajan N, Yona G: Automatic prediction of protein domains from sequence information using a hybrid learning system. Bioinformatics 2004, 20: 1335â€“60.
Gewehr JE, Zimmer R: SSEPDomain: protein domain prediction by alignment of secondary structure elements and profiles. Bioinformatics 2006, 22(2):181â€“187.
Dumontier M, Feldman R, Yao HJ, Hogue CWV: Armidillo: Doamin Boundary Prediction by Amino Acid Composition. J Mol Biol 2005, 350: 1061â€“1073.
Suyama M, Ohara O: DomCut: prediction of interdomain linker regions in amino acid sequences. Bioinformatics 2003, 19(5):673â€“674.
Sim J, Kim SY, Lee J: PRODO: Prediction of Protein Domain Boundaries using Neural Networks. Proteins 2005, 59: 627â€“632.
Sikder AR, Zomaya AY: Improving the performance of DomainDiscovery of protein domain boundary assignment using interdomain linker index. BMC Bioinformatics 2006, 7(Suppl 5):S6.
Dumontier M, Feldman R, Yao HJ, Hogue CWV: Armidillo: Doamin Boundary Prediction by Amino Acid Composition. J Mol Biol 2005, 350: 1061â€“1073.
Joshi RR: A Decade of Computing to Traverse the Labyrinth of Protein Domains. Current Bioinformatics 2007, 2: 113â€“131.
Altschul SF, Madden TL, Schaffer AA, Zhang J, Zhang Z, Miller W, Lipman DJ: Gapped BLAST and PSIBLAST: a new generation of protein database search programs. Nucleic Acids Research 1997, 25(17):3389â€“3402.
Gokhale RS, Khosla C: Role of linkers in communication between protein modules. Curr Opin Chem Biol 2000, 4: 22â€“27.
Civera C, Simon B, Stier G, Sattler M, Macias MJ: Structure and dynamics of the human pleckstrin DEP domain: distinct molecular features of a novel DEP domain subfamily. Proteins: Struct Funct Genet 2005, 58: 354â€“366.
Robinson CR, Sauer RT: Optimizing the stability of singlechain proteins by linker length and composition mutagenesis. Proc Natl Acad Sci USA 1998, 95: 5929â€“5934.
Van Leeuwen HC, Strating MJ, Rensen M, de Laat W, van der Vliet PC: Linker length and composition influence the flexibility of Oct1 DNA binding. EMBO J 1997, 16: 2043â€“2053.
Lehtinen MJ, Meri S, Jokiranta TS: Interdomain contact regions and angles between adjacent short consensus repeat domains. J Mol Biol 2004, 344: 1385â€“1396.
Liu J, Rost B: Sequencebased prediction of protein domains. Nucleic Acids. Research 2004, 32(12):3522â€“3530.
Holland TA, Veretnik S, Shindyalov IN, Bourne PE: Partitioning Protein Structures into Domains: Why Is It so Difficult? J Mol Biol 2006, 361(3):562â€“590.
Chen J, Chaudhari NS: Bidirectional SegmentedMemory Recurrent Neural Network for Protein Secondary Structure Prediction. Soft Computing Journal 2005, 10: 315â€“324.
Ceroni A, Frasconi P: On the role of longrange dependencies in learning protein secondary structure. IEEE Proc on Neural Networks 2004, 3: 1899â€“1904.
Pearl FMG, Lee D, Bray JE, Sillitoe I, Todd AE, Harrison AP, Thornton JM, Orengo CA: Assigning genomic sequences to CATH. Nucleic Acids Research 2000, 28(1):277â€“282.
Andreeva A, Howorth D, Brenner SE, Hubbard TJ, Chothia C, Murzin AG: SCOP database in 2004: refinements integrate structure and sequence family data. Nucleic Acids Res 2004, (32 Database):D226D229.
Berman HM, Westbrook J, Feng Z, Gilliland G, Bhat TN, Weissig H, Shindylov IN, Bourne PE: The Protein Data Bank. Nucleic Acids Research 2000, 28: 235â€“242.
Pollastri G, Baldi P, Fariselli P, Casadio R: Prediction of coordination number and relative solvent accessibility in proteins. Proteins 2002, 47: 142â€“153.
Pollastri G, Przybylski D, Rost B, Baldi P: Improving the prediction of protein secondary structure in three and eight classes using recurrent neural networks and profiles. Proteins 2002, 47: 228â€“235.
Vieira A, Oliva B: Protein Loop Classification Using Artificial Neural Networks. LNBI 2005, 3594: 222â€“225.
Melo JCB, Cavalcanti GDC, Guimaraes KS: PCA Feature Extraction for Protein Structure Prediction. IEEE Proc of the 2003 International Joint Conference on Neural Networks, Oregon, USA 2003.
Jain AK, Chandrasekaran B: Dimensionality and sample size considerations in pattern recognition in practice. 1982, 2: 835â€“855.
Freund Y, Schapire RE: Experiments with a new boosting algorithm. Machine Learning: Proceedings of the Thirteenth International Conference 1996, 148â€“156.
Schapire RE: A brief introduction to boosting. Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence 1999, 1401â€“1406.
Prompramote S, Chen Y, Chen Y: Machine Learning in Bioinformatics. In Bioinformatics Technologies. SpringerVerlag Berlin Heidelberg, Germany; 2005:117â€“153.
Lee CC, Chung PC, Tsai JR, Chang CI: Robust radial basis function neural network. IEEE Trans Syst Man Cybern, Part B, Cybern 1999, 29: 674â€“685.
Jang JSR, Sun CT, Mizutani E: NeuroFuzzy and Soft Computing. PrenticeHall, New Jersey; 1997.
Acknowledgements
We are grateful to Professor Philip Bourne and Dr Stella Veretnik of University of California San Diego, CA, USA for providing us Benchmark_2 Dataset. This work is partially supported by an Australian Research Council Grant number DP0667266 (ARS).
This article has been published as part of BMC Bioinformatics Volume 9 Supplement 1, 2008: Asia Pacific Bioinformatics Network (APBioNet) Sixth International Conference on Bioinformatics (InCoB2007). The full contents of the supplement are available online at http://www.biomedcentral.com/14712105/9?issue=S1.
Author information
Authors and Affiliations
Corresponding authors
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors' contributions
PDY developed and implemented the new model (IGRN) and drafted the manuscript. ARS prepared datasets, programming in C++, Shell and Perl scripts, and interpreted the results with PDY. BBZ and AYZ edited the manuscript and introduced the problem initially.
Paul D Yoo, Abdur R Sikder contributed equally to this work.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Yoo, P.D., Sikder, A.R., Zhou, B.B. et al. Improved general regression network for protein domain boundary prediction. BMC Bioinformatics 9 (Suppl 1), S12 (2008). https://doi.org/10.1186/147121059S1S12
Published:
DOI: https://doi.org/10.1186/147121059S1S12