Skip to main content

Predicting Class II MHC-Peptide binding: a kernel based approach using similarity scores



Modelling the interaction between potentially antigenic peptides and Major Histocompatibility Complex (MHC) molecules is a key step in identifying potential T-cell epitopes. For Class II MHC alleles, the binding groove is open at both ends, causing ambiguity in the positional alignment between the groove and peptide, as well as creating uncertainty as to what parts of the peptide interact with the MHC. Moreover, the antigenic peptides have variable lengths, making naive modelling methods difficult to apply. This paper introduces a kernel method that can handle variable length peptides effectively by quantifying similarities between peptide sequences and integrating these into the kernel.


The kernel approach presented here shows increased prediction accuracy with a significantly higher number of true positives and negatives on multiple MHC class II alleles, when testing data sets from MHCPEP [1], MCHBN [2], and MHCBench [3]. Evaluation by cross validation, when segregating binders and non-binders, produced an average of 0.824 AROC for the MHCBench data sets (up from 0.756), and an average of 0.96 AROC for multiple alleles of the MHCPEP database.


The method improves performance over existing state-of-the-art methods of MHC class II peptide binding predictions by using a custom, knowledge-based representation of peptides. Similarity scores, in contrast to a fixed-length, pocket-specific representation of amino acids, provide a flexible and powerful way of modelling MHC binding, and can easily be applied to other dynamic sequence problems.


Major Histocompatibility Complexes (MHC) bind short peptides derived from antigens and present them on the cell surface for inspection by T-cells. The binding mechanism appears to be the most selective step in the recognition of T-cell epitopes. The molecular mechanisms underlying this selectivity are still debated [4], but a crucial factor is the complementarity between amino acids in the antigen peptide and the MHC binding pocket [5]. Successfully modelling the behaviour exhibited by MHCs can be used to pre-select candidate peptides, which, in turn, can limit the practical work involved and facilitate the search for new vaccines.

MHC alleles are grouped according to their structure. For class I MHC alleles, the binding groove is closed at both ends, making it possible to predict exactly which residues are positioned in the binding groove. For Class II MHC molecules, the binding groove is open at both ends and peptides which bind class II alleles are generally longer than those which bind class I MHCs, typically 9 to 25 residues. Moreover, the grooves of MHC Class II alleles will only accommodate 9 to 11 residues of the target peptide [6]. Thus class II peptides have the potential to bind to the MHC groove in one of several registers (potential alignments between groove and antigenic peptide).

Interaction, within the groove, between MHC and peptide side chains is generally considered the principal determinant of binding affinity [7]. However, for MHC Class II type alleles, a recent study speculates that binding may not be completely deterministic, and that the same peptide can have multiple possible binding cores [8]. Moreover, several studies have shown that the binding core is, indeed, not the only factor; residues outside the binding groove (flanking residues) can also interact with the MHC molecule and influence binding [914]. Hence, this creates additional complexity in determining which residues are involved in the interaction, and suggests that a suitable method must include a full-length representation of the peptide.

Numerous methods have been applied to the problem of predicting MHC binding. Prediction of MHC class I binding has been very successful, reporting prediction accuracies of up to 95% (e.g. [15]). Attempts at predicting class II MHC binding show significantly lower accuracies, although many efforts using both traditional and novel approaches have been applied, some demonstrating inspirational progress.

In recent years, efficient pattern recognition methods have been applied to the class II problem, such as Artificial Neural Networks [7, 16, 17] and Support Vector Machines [18, 19]. However, these methods are based on inductive learning and require fixed-size representations to perform attribute-by-attribute comparisons of input variables. A typical approach for such methods is to first estimate (or input) a binding core, and subsequently predict the binding affinity of an unknown peptide based on the estimated core (typically a nonamer). This 2-step process is convenient from a mathematical modelling perspective, because it restricts the prediction task to a fixed-length formulation (9-mers) and thus avoids the problem of handling variable length peptides. The subsequent conversion of the 9-mer amino acid representation into a numerical representation is achieved by using either a binary positional system with 20 inputs per amino acid [17, 18, 2022], or by using amino acid properties [23, 24]. The results are fixed-length, high-dimensional, input vectors used for training the model (up to 180 dimensions in the case of the binary positional system).

Approaches for solving the dynamic nature of the prediction problem, and which can handle the variability in peptide lengths, have shown promising prediction qualities. Methods include an iterative "meta-search" algorithm [20], an iterative Partial Least Squares method [22], Hidden Markov Models [16, 25], an Ant Colony search [26], and a Gibbs sampling algorithm [27]. Some of these novel approaches have produced remarkable results, significantly outperforming conventional approaches.

The method presented here aims to combine the advantages of the two approaches: It utilises an efficient fixed-length discriminative method, but is still able to handle variable length peptides. This is achieved by applying a customised kernel.

Kernel methods have become popular thanks to Support Vector Machines (SVMs), originally introduced by Vapnik [28]. They have been applied to multiple bioinformatical problems, and have shown excellent performance using real-world data sets (see [29] for examples). In its basic form, a single SVM is a binary classifier which learns a decision boundary between two classes (e.g. binders and non-binders) in some input space (e.g. vectors with some amino acid representation). To find a decision boundary between two classes, an SVM attempts to maximise the margin between the classes, and choose a linear separation in a feature space. A function called the kernel function K (x i , x) is used to project the data from input space to feature space, and if this projection is non-linear it allows for non-linear decision boundaries. The effectiveness of SVMs is due to two factors: a) the principle of maximising margins (structural as opposed to empirical risk minimisation) and b) using the kernel trick to extend linear methods so that they can address non-linear problems. Details of the SVM formulation have been described thoroughly in many books and publications (e.g. [30]).

An advantage of kernel methods, which render them particularly suited for problems in computational biology, is the ability to customise the kernel. The kernel can be seen as a distance measure between two samples, e.g. in the case of a linear kernel the Euclidean distance between two samples. A custom kernel can be used to define explicitly a distance measure between two samples, and thus knowledge-based kernels can be designed to process variable length data and convert samples into fixed-length representations needed for direct comparisons. For sequences of proteins, it can be used to define similarity measures between pairs of sequences (proteins, peptide strings, etc). Methods utilising such direct kernel functions have lead to significant improvements in performance on classical bioinformatical problems, such as remote homology detection [31, 32] and protein classification [33].

In this paper, we present a kernel method based on the direct kernel function of [32], which we have adapted to the problem of predicting MHC binding. The Local Alignment Kernel is a kernel quantifying the similarity between a pair of protein sequences by taking into account all possible optimal alignment scores between all possible sub-sequences.


Using several sets of data (see Table 1), a method for the prediction of class II epitopes was developed and subsequently optimised. Initially, the effect on accuracy of varying the two parameters of the model was explored; these include a regulatory parameter β and a substitution matrix S (·), which are both described in detail below. Tests were then run to compare the performance of this kernel approach with existing prediction methods.

Table 1 Overview of data sets

Optimising the β-parameter

The kernel is based on similarity scores between pairs of peptides. For each pair, a similarity score is composed of multiple sub-scores based on alignments between pairs of sub-sequences. The model parameter β regulates the relative influence that each sub-score will have on the cumulative score. In turn, it enables adjustment of the importance of sub-optimal alignments.

Experiments were undertaken to evaluate the effect on performance of varying the β-parameter using a simple test set. The MHCBench Set 4b was chosen for this purpose; it contains experimentally verified binders and non-binders of HLA-DRB1*0401. It consists of only natural peptides and an equal number of binders and non-binders (292 of each), which makes it well-suited for model testing.

The BLOSUM62 substitution matrix was used for S (·). This matrix is generally considered to be a good matrix for modelling evolutionary problems [34]. 10-fold CV was used to evaluate performance, and a rough search for a good β-parameter was undertaken. The effect on performance of varying β can be seen in Figure 1, which shows that the SKM is capable of distinguishing well between binders and non-binders. The best accuracy is 74.7% at β = 0.025. The best performance in terms of AROC is 0.827 at β = 0.035. Generally, the best results for most measures are found for β-values between 0.02 and 0.04. Higher β values of 0.2 to 5.0 were also evaluated, but as β becomes larger performance degrades for all measures.

Figure 1

Evaluating performance for varying the β-parameter using 10-fold CV. Graphs are plotted for accuracy (proportion of correct predictions), sensitivity (proportion of false negatives), specificity (proportion of true positives), and AROC (area under receiver operating characteristic curve).

Bootstrapping using case resampling [35] was performed to analyse the variance in results. 100 repetitions were undertaken, with data set sizes of 584. At a β-value of 0.025, which produced the best accuracy in the tests referred to above, the average bootstrapping accuracy was 73.1% with a standard deviation of 2.2%. This degree of variance was found throughout our experiments.

Interestingly, a low β indicates that the best solution is found when sub-optimal alignments have a large influence, as seen by the mathematical formulation below; when lowering the β-value, the relative contributions from scores of various alignments are evened out. This tendency was observed throughout the remaining experiments, with most "optimal" β-values being below 0.1. Interestingly, the same observation regarding the positive influence of sub-optimal alignments was also reported in Nielsen et al [27] using a very different method.

Selecting the substitution matrix

A substitution matrix was used in the calculation of the Smith-Waterman score to evaluate similarities between amino acids (see Eq. 0.3). The BLOSUM62 matrix initially used is regarded as a good descriptor for evolutionary problems [34]. However, using an alternative substitution matrix could prove more effective.

The AAIndex database [23] contains a large collection of substitution matrices produced during the last three decades. The matrices are based on numerous different measures, such as physicochemical properties and structural differences. An extensive search among the substitution matrices was conducted. Each substitution matrix was used instead of the BLOSUM62 matrix as above. Due to the scale of the experiment, only a crude search using 5 values of β was evaluated per substitution matrix. From the 83 matrices, the ten best performing substitution matrices with regard to AROC scores were retested with a refined search for the best value of β. The 3 best performing substitution matrices from this experiment are shown in Table 2.

Table 2 Results from varying substitution matrices

As can be seen from the table, the three matrices have very similar AROC values. The best performance was produced by a recently developed substitution matrix SM_THREADER_NORM, which is based on molecular mechanics force fields. [36] suggest that force fields can provide more reliable mutation matrices because of the incorporation of natural weighting of different physical contributions. Interestingly, the BLOSUM62 matrix is among the best three matrices out of 83. This suggests that the evolutionary rationale behind BLOSUM62 is also appropriate for MHC peptide similarity or that the chemical similarities underlying protein evolution also underlie peptide selectivity by the MHC. Such conjecture is supported in part by the fact that the SM_THREADER_NORM is also placed in the same family of substitution matrices when assessing the magnitude of distances between matrices [36]. In the following experiments, the SM_THREADER_NORM is used.

Performance on HLA-DRB1*0101 and HLA-DRB1*0301

Two MHC Class II alleles from the MHCBN database [2] were evaluated. The MHCBN database contains 475 binders and 105 non-binders for HLA-DRB1*0101 and for HLA-DRB1*0301 contains 219 binders and 150 non-binders. Duplicates and peptides with 75% or more Alanines were removed. 5-fold cross-validation was undertaken (5 fold CV used instead of 10-fold CV for comparison with [21]), and a crude optimisation of the β-parameter was performed as described in the methods section. Results are shown in Table 3. The table reports comparison results from two other methods, a linear programming model "LP_top2" [21] and TEPITOPE, a quantitative matrix method [37].

Table 3 Results on HLA-DRB1*0101 and HLA-DRB1*0301

The results in Table 3 show that the SKM method performs significantly better than LP_top2 and TEPITOPE. The relative improvements in AROC scores are 8% and 41%, and the improvements in AOVER-ROC are 37% to 60%. Other methods have also been evaluated on the data sets. In [25], results on two methods using Hidden Markov models combined with successive state splitting are reported. The best 10-fold CV results were 0.85 (S-HMM1) and 0.89 (S-HMM2), which are close to but still lower than the 0.91 AROC of SKM using 5-fold CV; using 10-fold CV the SKM performance increases to 0.93 AROC.

Performance on allele HLA-DRB1*0401

MHCBench [3] contains 8 data sets of binders and non-binders for HLA-DRB1*0401. Again, the SKM was evaluated using 10-fold cross validation, and a crude search for optimal values of β was performed for each set. AROC performance from all 8 sets are reported in Table 4, which also includes results of PERUN, a method based on TEPITOPE [38], Neural Networks [7], Gibbs Sampler, a method based on Metropolis sampling [27], and LP_top2, a linear programming method [21].

Table 4 Comparison of AROC values on HLA-DRB1*0401 data sets from MHCBench

The results on Table 4 show that the SKM is significantly better on 6 out of 8 benchmark sets (Set 1 to Set 4b). On sets 5a and 5b, another method, LP_top2, scored the best results (0.815 AROC and 0.859 AROC, respectively). The relative lower accuracies of the SKM method on these two data sets may be due to the small training set sizes (117 and 85); in training, the SKM method selected nearly all training samples as Support Vectors, indicating there may not be sufficient samples to properly describe the model space.

Alleles from MHCPEP

All MHC Class II alleles from the MHCPEP database were evaluated. This database contains only binders, with binding strengths graded from low to high. An extensive number of alleles were tested, making it hard to obtain known non-binders for the sets. One approach is to use binders from other alleles as non-binders for the allele of interest. However, more than 10% of peptides were found to bind two or more alleles, which would generate a large amount of noise and uncertainty in predictions if used as non-binders. Instead, the non-binders were generated randomly in order to have the same length-distribution as the set of binders. Yewdell et al estimated that only one in 100 to 200 peptides will bind to an average allele [5], which makes this approach a reasonable approximation. Moreover, generating an equal amount of binders and non-binders creates a balanced data set well suited for computational experiments.

For each allele, the data set was extracted from MHCPEP and evaluated using 10-fold CV. As in previous experiments, a crude search for the best value of β was conducted. The results can be seen in Table 5. The SKM is able to model multiple MHC Class II alleles well, with an average AROC of well over 0.9 in all cases except for HLA-DR10, which is a data set containing only 6 binders. Overall, the average performance (0.967 AROC) is competitive (as good as or better) compared with literature results.

Table 5 Results of SKM on multiple MHC Class II alleles from MHCPEP

In [17], internal test sets of *0101 and *0301 extracted from MHCPEP were predicted with 0.91 AROC and 0.88 AROC, respectively. In comparison, the corresponding results for the SKM are 0.966 and 0.990. In [20], an iterative stepwise discriminant analysis was run on 400 HLA-DR1 high and moderate binders from MHCPEP and 743 non-binders. Classification accuracy using Jack-knife cross-validation was 91.4%. Here, the overall classification accuracy is slightly better at 92.2%.

In [16], a fuzzy neural network, combined with 3 amino acid property descriptors, was used to separate high, moderate and low binders from non-binders of HLA-DRB1*0401. Of the 321 binders and 312 non-binders collected for the allele, the highest performance was on strong (high affinity) binders vs. non-binders with an accuracy of 0.94 AROC. However, for moderate and low affinity binders, their results were 0.93 and 0.88, respectively. For the SKM, the average AROC value for all binders vs. non-binders is 0.952.

In [39], an ensemble classifier based on Support Vector Machines with a representation using QSAR descriptors achieved 0.917 AROC on a data set consisting of 9-mers of HLA-DR4. As in this study, non-binders were generated randomly. The performance of the SKM (0.972 AROC) on HLA-DR4 is significantly higher.


The proposed kernel method is shown to provide excellent discrimination between binders and non-binders for multiple alleles. It is able to model the dynamic MHC class II problem, and produce results that compare favourably with previously published results. The reason for the good performance may be due to several factors, and it is important to identify which of these are the most significant. The main focus of modelling was to consider the full length of peptides, as studies have shown that peptides outside the binding core can influence binding affinity [9]. Avoiding estimation of binding cores eliminates the potential for using faulty alignments, which can lead to increased model noise and, in turn, lower accuracy.

The use of similarity scores is a significant conceptual change in peptide evaluation, quantifying the overall similarity between peptides and interrelations between residues. This concept contrasts to the fixed-length representation (using binary positional system or amino acid properties) which enforces a direct pocket-to-pocket comparison of residues. Most static pattern recognition methods consider each input property to be a separate and independent entity, which is clearly not the case for a peptide string. Instead, higher order interactions within the peptide may also make a significant contribution to the modulation of affinity. Hattotuwagama et al. showed that the motif-dependence of Class II peptides is even weaker than that of class I epitopes [40]. Modelling such subtle effects, also seen in X-ray structures, are beyond the scope of much existing prediction technology. This change in concept could be a significant reason for the improved performance. Incorporating sub-optimal alignments into similarity scores have certainly contributed to an observed improvement, where low values of β produced the best performance.

Kernel methods, such as Support Vector Machines, have previously been shown to work well on biological problems, particularly when custom engineered kernels are used [31, 33]. The SVM itself and the training principle of structural risk minimisation may have contributed to enhanced performance. However, simply applying SVMs to the MHC problem using aligned and truncated peptides (9-mers), in combination with a binary representation of amino acids similar to [18], did not produce promising results in initial experiments; custom kernels must be used to take full advantage of the kernel machines' excellent capacity for generalisation. Another advantage of using kernel methods is the ability to choose a kernel method independent of the choice of kernel itself. Thus, a kernel can readily be combined with a range of different kernel methods. This is useful when certain properties of the predictor are desired; e.g. some kernel methods can handle large-scale data sets while others allow for probabilistic interpretation of outputs.

Naturally, the proposed method is not without it's disadvantages. Firstly, the method is purely data-driven, in the sense that it relies solely on information derived from peptide data sets and thus does not consider MHC allele-specific structural information about the binding groove. While this may be seen as an advantage, since it keeps assumptions to a minimum, potentially important information is not considered, such as a specific pockets' preference for certain amino acids. Secondly, the method does not attempt to estimate alignments, which may be of interest, and finally, computational complexity and run-time speed could also be an issue for large scale testing. Calculating the kernel is time-consuming even with an efficient implementation; it cannot currently handle more than a few thousand samples before run-time becomes prohibitive.

Many potential improvements are possible that could either improve classification accuracy or provide more informative results. More advanced kernels could be developed: by increasing the importance of similarity scores of certain sized windows (e.g. length of 9) and subsequently weighting each residue in the window according to known binding motifs (e.g. [41]). This would have the advantage of incorporating allele-specific information into the method. Other improvements include modifying the kernel method to improve training or classification speeds, and developing new substitution matrices specific to the MHC domain similar to that undertaken for trans-membrane proteins [42, 43]. Finally, the binary classification could be extended to a multi-class problem (separating non-binders from low, medium and high affinity binders), or directly predicting binding affinity by kernel regression, as Lui et al [44] has done for class I.


The combination of a complex similarity score and an efficient kernel method are shown here to be a powerful tool for predicting MHC class II peptide binding affinity. The principle of using kernels to define similarities between sequences explicitly is a simple, yet flexible and powerful, way of modelling sequence data, and can readily be extended to address a variety of immunological and other biological problems.


Kernel engineering and string kernels

The mathematical formulation of kernel machines are described in details in books and publications (e.g. [30]). The kernel function K (x i , x) is the core of any kernel method, and can be used to incorporate a-priori knowledge of the problem into the model. A kernel function corresponds to a measurement of similarity or difference between any pair of samples (e.g. the linear kernel is a measure of the Euclidean distance between samples). However, kernel functions do not need to measure pair-wise similarities through a dot-product of vector representations. Instead, explicit measures of similarities between samples can be used; such as similarities between amino acid strings: string kernels. This enables the full length of each peptide to be incorporated into the model, including information known to be hidden in flanking residues [1014, 45].

Local alignment kernels

The principle of local alignments has been shown to provide a powerful approach to detecting relationships between sequences, using the optimal local alignment via the Smith-Waterman algorithm [46] and it's efficient (PSI-)BLAST approximations [47]. Therefore, we utilise a complex kernel comprising several sub-kernels based on the Smith-Waterman algorithm [32]. On a protein homology detection problem, this approach was found to significantly outperform scores based solely on optimal alignments [31].

Local Alignment Kernels are convolution kernels [48] consisting of a number of simple sub-kernels:

K 1 K 2 ... K p ( x , y ) = x = x 1 ... x p , y = y 1 ... y p K 1 ( x 1 , y 1 ) ... K p ( x p , y p ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaafaqaaeGabaaabaGaem4saS0aaSbaaSqaaiabigdaXaqabaGccqGHflY1cqWGlbWsdaWgaaWcbaGaeGOmaidabeaakiabgwSixlabc6caUiabc6caUiabc6caUiabgwSixlabdUealnaaBaaaleaacqWGWbaCaeqaaOGaeiikaGIaemiEaGNaeiilaWIaemyEaKNaeiykaKIaeyypa0dabaWaaabuaeaacqWGlbWsdaWgaaWcbaGaeGymaedabeaakiabcIcaOiabdIha4naaBaaaleaacqaIXaqmaeqaaOGaeiilaWIaemyEaK3aaSbaaSqaaiabigdaXaqabaGccqGGPaqkcqGHflY1cqGGUaGlcqGGUaGlcqGGUaGlcqGHflY1cqWGlbWsdaWgaaWcbaGaemiCaahabeaakiabcIcaOiabdIha4naaBaaaleaacqWGWbaCaeqaaOGaeiilaWIaemyEaK3aaSbaaSqaaiabdchaWbqabaGccqGGPaqkaSqaaiabdIha4jabg2da9iabdIha4naaBaaameaacqaIXaqmaeqaaSGaeiOla4IaeiOla4IaeiOla4IaemiEaG3aaSbaaWqaaiabdchaWbqabaWccqGGSaalcqWG5bqEcqGH9aqpcqWG5bqEdaWgaaadbaGaeGymaedabeaaliabc6caUiabc6caUiabc6caUiabdMha5naaBaaameaacqWGWbaCaeqaaaWcbeqdcqGHris5aaaaaaa@79C6@

Eq. 0.1

Where the components K1...Kp consists of three different kernels: (a) A constant kernel Kconst, (b) a kernel for measuring the difference between aligned letters Kalign, and (c) a kernel for penalizing gaps Kgap:

K c o n s t ( x , y ) = 1 K a l i g n ( x , y ) = { 0 , i f | x | 1 o r | y | 1 e β S ( x , y ) o t h e r w i s e K g a p ( x , y ) = e β ( g ( | x | ) + g ( | y | ) ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaafaqaaeWabaaabaGaem4saS0aaSbaaSqaaiabdogaJjabd+gaVjabd6gaUjabdohaZjabdsha0bqabaGccqGGOaakcqWG4baEcqGGSaalcqWG5bqEcqGGPaqkcqGH9aqpcqaIXaqmaeaacqWGlbWsdaWgaaWcbaGaemyyaeMaemiBaWMaemyAaKMaem4zaCMaemOBa4gabeaakiabcIcaOiabdIha4jabcYcaSiabdMha5jabcMcaPiabg2da9maaceqabaqbaeqabiGaaaqaaiabicdaWaqaaiabcYcaSiabdMgaPjabdAgaMjabcYha8jabdIha4jabcYha8jabgcMi5kabigdaXiabbccaGiabd+gaVjabdkhaYjabcYha8jabdMha5jabcYha8jabgcMi5kabigdaXaqaaiabdwgaLnaaCaaaleqabaacciGae8NSdiMaey4fIOIaem4uamLaeiikaGIaemiEaGNaeiilaWIaemyEaKNaeiykaKcaaaGcbaGaem4Ba8MaemiDaqNaemiAaGMaemyzauMaemOCaiNaem4DaCNaemyAaKMaem4CamNaemyzaugaaaGaay5EaaaabaGaem4saS0aaSbaaSqaaiabdEgaNjabdggaHjabdchaWbqabaGccqGGOaakcqWG4baEcqGGSaalcqWG5bqEcqGGPaqkcqGH9aqpcqWGLbqzdaahaaWcbeqaaiab=j7aIjabcIcaOiabdEgaNjabcIcaOiabcYha8jabdIha4jabcYha8jabcMcaPiabgUcaRiabdEgaNjabcIcaOiabcYha8jabdMha5jabcYha8jabcMcaPiabcMcaPaaaaaaaaa@9ACD@

Eq. 0.2

where x and y are the amino acid sequences, S (x, y) the Smith-Waterman score, g (·) the gap penalty function, and β a scaling parameter to adjust importance of gaps and sub-optimal alignments.

The Smith-Waterman score SWS,g(π)is calculated as:

S S , g ( π ) = i = 1 | π | S ( x π 1 ( i ) , y π 2 ( i ) ) i = 1 | π | 1 [ g ( π 1 ( i + 1 ) π 1 ( i ) ) + g ( π 2 ( i + 1 ) π 2 ( i ) ) ] S W S , g ( π ) ( x , y ) = max π Π ( x , y ) S S , g ( π ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaafaqaaeGabaaabaGaem4uam1aaSbaaSqaaiabdofatjabcYcaSiabdEgaNjabcIcaOGGaciab=b8aWjabcMcaPaqabaGccqGH9aqpdaaeWbqaaiabdofatnaabmaabaGaemiEaG3aaSbaaSqaaiab=b8aWjabigdaXiabcIcaOiabdMgaPjabcMcaPaqabaGccqGGSaalcqWG5bqEdaWgaaWcbaGae8hWdaNaeGOmaiJaeiikaGIaemyAaKMaeiykaKcabeaaaOGaayjkaiaawMcaaaWcbaGaemyAaKMaeyypa0JaeGymaedabaGaeiiFaWNae8hWdaNaeiiFaWhaniabggHiLdGccqGHsisldaaeWbqaaiabcUfaBjabdEgaNjabcIcaOiab=b8aWnaaBaaaleaacqaIXaqmaeqaaOGaeiikaGIaemyAaKMaey4kaSIaeGymaeJaeiykaKIaeyOeI0Iae8hWda3aaSbaaSqaaiabigdaXaqabaGccqGGOaakcqWGPbqAcqGGPaqkcqGGPaqkcqGHRaWkcqWGNbWzcqGGOaakcqWFapaCdaWgaaWcbaGaeGOmaidabeaakiabcIcaOiabdMgaPjabgUcaRiabigdaXiabcMcaPiabgkHiTiab=b8aWnaaBaaaleaacqaIYaGmaeqaaOGaeiikaGIaemyAaKMaeiykaKIaeiykaKIaeiyxa0faleaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqGG8baFcqWFapaCcqGG8baFcqGHsislcqaIXaqma0GaeyyeIuoaaOqaaGqaciab+nfatjab+DfaxnaaBaaaleaacqGFtbWucqGFSaalcqGFNbWzieaacqqFOaakcqWFapaCcqqFPaqkaeqaaOGaeiikaGIaemiEaGNaeiilaWIaemyEaKNaeiykaKIaeyypa0ZaaCbeaeaacyGGTbqBcqGGHbqycqGG4baEaSqaaiab=b8aWjabgIGiolabfc6aqjabcIcaOiabdIha4jabcYcaSiabdMha5jabcMcaPaqabaGccqWGtbWudaWgaaWcbaGaem4uamLaeiilaWIaem4zaCMaeiikaGIae8hWdaNaeiykaKcabeaaaaaaaa@AD37@

Eq. 0.3

Where π is the alignment between two sequences x and y, S (·) is a substitution matrix and g (·) a gap penalty function.

The component sub-kernels are combined by convolution to represent a kernel for an alignment of length n. The Local Alignment score is the sum over all possible alignments in the sequence:

K ( n ) ( x , y ) = K c o n s t ( K a l i g n K g a p ) ( n 1 ) K a l i g n K c o n s t K L A ( x , y ) = i = 0 N K i ( x , y ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaafaqaaeGabaaabaGaem4saS0aaSbaaSqaaiabcIcaOiabd6gaUjabcMcaPaqabaGccqGGOaakcqWG4baEcqGGSaalcqWG5bqEcqGGPaqkcqGH9aqpcqWGlbWsdaWgaaWcbaGaem4yamMaem4Ba8MaemOBa4Maem4CamNaemiDaqhabeaakiabgwSixlabcIcaOiabdUealnaaBaaaleaacqWGHbqycqWGSbaBcqWGPbqAcqWGNbWzcqWGUbGBaeqaaOGaeyyXICTaem4saS0aaSbaaSqaaiabdEgaNjabdggaHjabdchaWbqabaGccqGGPaqkdaahaaWcbeqaaiabcIcaOiabd6gaUjabgkHiTiabigdaXiabcMcaPaaakiabgwSixlabdUealnaaBaaaleaacqWGHbqycqWGSbaBcqWGPbqAcqWGNbWzcqWGUbGBaeqaaOGaeyyXICTaem4saS0aaSbaaSqaaiabdogaJjabd+gaVjabd6gaUjabdohaZjabdsha0bqabaaakeaacqWGlbWsdaWgaaWcbaGaemitaWKaemyqaeeabeaakiabcIcaOiabdIha4jabcYcaSiabdMha5jabcMcaPiabg2da9maaqahabaGaem4saS0aaSbaaSqaaiabdMgaPbqabaGccqGGOaakcqWG4baEcqGGSaalcqWG5bqEcqGGPaqkaSqaaiabdMgaPjabg2da9iabicdaWaqaaiabd6eaobqdcqGHris5aaaaaaa@86DE@

Eq. 0.4

where N is the number of possible alignments.

The above formulation results in a computational complexity exponential with |x| and |y|, and is thus not a feasible solution for this problem. Hence, a dynamic programming algorithm by [32] is used, which is a slight modification of the Smith-Waterman algorithm [46]. The kernel computation is done in O (n2·|x|·|y|), where n is the number of samples, and |x| and |y| are the lengths of the peptide strings. In the context of MHC-peptide binding, the gap penalty term must be maximised since gaps are not possible within bound peptides.

Calculating the sub-kernels requires the following two parameters: the substitution matrix S (·) in Eq. 0.3, and the β parameter in Eq. 0.2. S (·) quantifies a similarity between pairs of amino acids, and is a well-known term in bioinformatics with numerous substitution matrices designed from evolutionary, physicochemical or structural properties (a list can be found at [23]). The β parameter regulates the effect of individual contributions from alignments, and allows adjustment of the relative importance between low- and high scoring alignments. When β is low, the model will increase the importance of low scoring (sub-optimal) alignments to quantify the similarity between sequences. Similarly, as β → ∞ the contributions from sub-optimal values is reduced.

All kernels must be symmetric and positive semi-definite. Some values of β and substitution matrices S in Eq. 0.2 resulted in invalid kernels, and caused convergence problems. A trick used in [32] subtracts the smallest negative Eigenvalue from the diagonal of the kernel to ensure kernels are positive semi-definite. Symmetry of the resulting kernel, K LA , is guaranteed as long as substitution matrix S (·) is symmetric.

Data set

Multiple data sets were used in the experiments. Eight benchmark data sets with samples of known binders and non-binders of the HLA-DRB1*0401 allele were taken from MHCBench [3]. Within the data sets, peptide strings are assigned binding strengths of level 0 (non-binders) to 4 (strong binders), and are collected from multiple sources, mainly MHCPEP [1]. The sets are derived from the same base set of peptides, created with varying levels of curation. Set 1 includes all peptides whereas Set5b is a homology reduced set containing only natural peptides. Data set sizes range from 85 to 1017 samples with peptide lengths of 9 to 33. As these sets have been used in many published experiments [21, 26, 27], we use them in preference to alternatives.

In addition to the 8 sets, two data sets from specific alleles HLA-DRB1*0101 and HLA-DRB1*0301 were taken from MHCBN [2]. The data sets separate peptides into binders (low, moderate, and high), and non-binders (peptides having IC50 values of more than 50,000 nm). Finally, binders from multiple alleles from the MHCPEP database [1] were used. In the database, peptides are labelled as having low, moderate or high binding affinity.

Experimental setup

MATLAB [49] was used as the testing environment, with assistance of the SPIDER toolbox [50]. The SKM was calculated with a C++ Mex implementation based on [32]. Testing machine was an Intel Pentium M 1.4GHz.

In all experiments, samples were randomly permuted and subsequently evaluated using N-fold cross validation (CV). Targets yi were divided into binders and non-binders; yi {-1, 1}. For each left-out fold, a model was trained on the remaining folds to separate samples into binders and non-binders. The trained model was then evaluated on the left out fold. In addition, a rough search for good values of β was performed by performing a full CV test for each value of β (typically evaluating 10–15 different values of β). CV is well-suited for model assessment of small data sets. However, some studies report the possibility of high variance in results using CV (e.g. [51]).

For assessing performance, several measures were used: Overall prediction accuracy, sensitivity, specificity, Matthew's Correlation Coefficient (MCC), area under receiver operating characteristic curve (AROC), and the AROC score up to the first 50 false positives (AROC50). Finally, in cases where AROC scores were close to 1.0 (perfect classification), the error term of area over the ROC curve, AOVER_ROC, was used as well.

For experiments, data sets were curated by removing duplicates as well as unnatural peptides with more than 75% Alanine [21, 27]. An overview of the data sets can be seen in Table 1.


  1. 1.

    Brusic V, Rudy G, Harrison LC: MHCPEP, a database of MHC-binding peptides: update 1997. Nucleic Acids Res 1998, 26(1):368–371. 10.1093/nar/26.1.368

    PubMed Central  CAS  Article  PubMed  Google Scholar 

  2. 2.

    Bhasin M, Singh H, Raghava GP: MHCBN: a comprehensive database of MHC binding and non-binding peptides. Bioinformatics 2003, 19(5):665–666. 10.1093/bioinformatics/btg055

    CAS  Article  PubMed  Google Scholar 

  3. 3.

    Raghava GP: MHCBench: Evaluation of MHC Binding Peptide Prediction Algorithms. 2001.

    Google Scholar 

  4. 4.

    Rhodes DA, Trowsdale J: Genetics and molecular genetics of the MHC. Rev Immunogenet 1999, 1(1):21–31.

    CAS  PubMed  Google Scholar 

  5. 5.

    Yewdell JW, Bennink JR: Immunodominance in major histocompatibility complex class I-restricted T lymphocyte responses. Annu Rev Immunol 1999, 17: 51–88. 10.1146/annurev.immunol.17.1.51

    CAS  Article  PubMed  Google Scholar 

  6. 6.

    Kropshofer H, Max H, Halder T, Kalbus M, Muller CA, Kalbacher H: Self-peptides from four HLA-DR alleles share hydrophobic anchor residues near the NH2-terminal including proline as a stop signal for trimming. J Immunol 1993, 151(9):4732–4742.

    CAS  PubMed  Google Scholar 

  7. 7.

    Brusic V, Rudy G, Honeyman G, Hammer J, Harrison L: Prediction of MHC class II-binding peptides using an evolutionary algorithm and artificial neural network. Bioinformatics 1998, 14(2):121–130. 10.1093/bioinformatics/14.2.121

    CAS  Article  PubMed  Google Scholar 

  8. 8.

    Tong JC, Zhang GL, Tan TW, August JT, Brusic V, Ranganathan S: Prediction of HLA-DQ3.2beta ligands: evidence of multiple registers in class II binding peptides. Bioinformatics 2006, 22(10):1232–1238. 10.1093/bioinformatics/btl071

    CAS  Article  PubMed  Google Scholar 

  9. 9.

    Zavala-Ruiz Z, Strug I, Anderson MW, Gorski J, Stern LJ: A polymorphic pocket at the P10 position contributes to peptide binding specificity in class II MHC proteins. Chem Biol 2004, 11(10):1395–1402. 10.1016/j.chembiol.2004.08.007

    CAS  Article  PubMed  Google Scholar 

  10. 10.

    Xia J, Siegel M, Bergseng E, Sollid LM, Khosla C: Inhibition of HLA-DQ2-mediated antigen presentation by analogues of a high affinity 33-residue peptide from alpha2-gliadin. J Am Chem Soc 2006, 128(6):1859–1867. 10.1021/ja056423o

    PubMed Central  CAS  Article  PubMed  Google Scholar 

  11. 11.

    Carson RT, Vignali KM, Woodland DL, Vignali DA: T cell receptor recognition of MHC class II-bound peptide flanking residues enhances immunogenicity and results in altered TCR V region usage. Immunity 1997, 7(3):387–399. 10.1016/S1074-7613(00)80360-X

    CAS  Article  PubMed  Google Scholar 

  12. 12.

    Bonomi G, Moschella F, Ombra MN, Del Pozzo G, Granier C, De Berardinis P, Guardiola J: Modulation of TCR recognition of MHC class II/peptide by processed remote N- and C-terminal epitope extensions. Hum Immunol 2000, 61(8):753–763. 10.1016/S0198-8859(00)00147-6

    CAS  Article  PubMed  Google Scholar 

  13. 13.

    Arnold PY, La Gruta NL, Miller T, Vignali KM, Adams PS, Woodland DL, Vignali DA: The majority of immunogenic epitopes generate CD4+ T cells that are dependent on MHC class II-bound peptide-flanking residues. J Immunol 2002, 169(2):739–749.

    CAS  Article  PubMed  Google Scholar 

  14. 14.

    Godkin AJ, Smith KJ, Willis A, Tejada-Simon MV, Zhang J, Elliott T, Hill AV: Naturally processed HLA class II peptides reveal highly conserved immunogenic flanking region sequence preferences that reflect antigen processing rather than peptide-MHC interactions. J Immunol 2001, 166(11):6720–6727.

    CAS  Article  PubMed  Google Scholar 

  15. 15.

    Donnes P, Elofsson A: Prediction of MHC class I binding peptides, using SVMHC. BMC Bioinformatics 2002, 3: 25. 10.1186/1471-2105-3-25

    PubMed Central  Article  PubMed  Google Scholar 

  16. 16.

    Noguchi H, Hanai T, Honda H, Harrison LC, Kobayashi T: Fuzzy neural network-based prediction of the motif for MHC class II binding peptides. J Biosci Bioeng 2001, 92(3):227–231. 10.1263/jbb.92.227

    CAS  Article  PubMed  Google Scholar 

  17. 17.

    Burden FR, Winkler DA: Predictive Bayesian neural network models of MHC class II peptide binding. J Mol Graph Model 2005, 23(6):481–489. 10.1016/j.jmgm.2005.03.001

    CAS  Article  PubMed  Google Scholar 

  18. 18.

    Bhasin M, Raghava GP: SVM based method for predicting HLA-DRB1*0401 binding peptides in an antigen sequence. Bioinformatics 2004, 20(3):421–423. 10.1093/bioinformatics/btg424

    CAS  Article  PubMed  Google Scholar 

  19. 19.

    Yang ZR, Johnson FC: Prediction of T-cell epitopes using biosupport vector machines. J Chem Inf Model 2005, 45(5):1424–1428. 10.1021/ci050004t

    CAS  Article  PubMed  Google Scholar 

  20. 20.

    Mallios RR: Predicting class II MHC/peptide multi-level binding with an iterative stepwise discriminant analysis meta-algorithm. Bioinformatics 2001, 17(10):942–948. 10.1093/bioinformatics/17.10.942

    CAS  Article  PubMed  Google Scholar 

  21. 21.

    Murugan N, Dai Y: Prediction of MHC class II binding peptides based on an iterative learning model. Immunome Res 2005, 1: 6. 10.1186/1745-7580-1-6

    PubMed Central  Article  PubMed  Google Scholar 

  22. 22.

    Doytchinova IA, Flower DR: Towards the in silico identification of class II restricted T-cell epitopes: a partial least squares iterative self-consistent algorithm for affinity prediction. Bioinformatics 2003, 19(17):2263–2270. 10.1093/bioinformatics/btg312

    CAS  Article  PubMed  Google Scholar 

  23. 23.

    Kawashima S, Ogata H, Kanehisa M: AAindex: Amino Acid Index Database. Nucleic Acids Res 1999, 27(1):368–369. 10.1093/nar/27.1.368

    PubMed Central  CAS  Article  PubMed  Google Scholar 

  24. 24.

    Guan P, Doytchinova IA, Walshe VA, Borrow P, Flower DR: Analysis of peptide-protein binding using amino acid descriptors: prediction and experimental verification for human histocompatibility complex HLA-A0201. J Med Chem 2005, 48(23):7418–7425. 10.1021/jm0505258

    CAS  Article  PubMed  Google Scholar 

  25. 25.

    Noguchi H, Kato R, Hanai T, Matsubara Y, Honda H, Brusic V, Kobayashi T: Hidden Markov model-based prediction of antigenic peptides that interact with MHC class II molecules. J Biosci Bioeng 2002, 94(3):264–270. 10.1263/jbb.94.264

    CAS  Article  PubMed  Google Scholar 

  26. 26.

    Karpenko O, Shi J, Dai Y: Prediction of MHC class II binders using the ant colony search strategy. Artif Intell Med 2005, 35(1–2):147–156. 10.1016/j.artmed.2005.02.002

    Article  PubMed  Google Scholar 

  27. 27.

    Nielsen M, Lundegaard C, Worning P, Hvid CS, Lamberth K, Buus S, Brunak S, Lund O: Improved prediction of MHC class I and class II epitopes using a novel Gibbs sampling approach. Bioinformatics 2004, 20(9):1388–1397. 10.1093/bioinformatics/bth100

    CAS  Article  PubMed  Google Scholar 

  28. 28.

    Vapnik VN: The nature of statistical learning theory. New York, Springer; 1995:xv, 188 p..

    Google Scholar 

  29. 29.

    Schölkopf B, Smola A, Müller KR: Nonlinear Component Analysis as a Kernel Eigenvalue Problem. Neural Computation 1998, 10: 1299–1319. 10.1162/089976698300017467

    Article  Google Scholar 

  30. 30.

    Schölkopf B, Smola AJ: Learning with kernels : support vector machines, regularization, optimization, and beyond. In Adaptive computation and machine learning. Cambridge, Mass., MIT Press; 2002:xviii, 626 p..

    Google Scholar 

  31. 31.

    Saigo H, Vert JP, Ueda N, Akutsu T: Protein homology detection using string alignment kernels. Bioinformatics 2004, 20(11):1682–1689. 10.1093/bioinformatics/bth141

    CAS  Article  PubMed  Google Scholar 

  32. 32.

    Vert JP, Akutsu T, Saigo H: Local Alignment Kernels for Biological Sequences. In Kernel Methods in Computational Biology. Edited by: Schölkopf, Tsuda, Vert . MIT Press; 2004.

    Google Scholar 

  33. 33.

    Kuang R, Ie E, Wang K, Wang K, Siddiqi M, Freund Y, Leslie C: Profile-based string kernels for remote homology detection and motif extraction. J Bioinform Comput Biol 2005, 3(3):527–550. 10.1142/S021972000500120X

    CAS  Article  PubMed  Google Scholar 

  34. 34.

    Henikoff S, Henikoff JG: Amino acid substitution matrices from protein blocks. Proc Natl Acad Sci U S A 1992, 89(22):10915–10919. 10.1073/pnas.89.22.10915

    PubMed Central  CAS  Article  PubMed  Google Scholar 

  35. 35.

    Efron B: The jackknife, the bootstrap, and other resampling plans. In CBMS-NSF regional conference series in applied mathematics ; 38. Philadelphia, Pa., Society for Industrial and Applied Mathematics; 1982:vii, 92 p..

    Google Scholar 

  36. 36.

    Dosztanyi Z, Torda AE: Amino acid similarity matrices based on force fields. Bioinformatics 2001, 17(8):686–699. 10.1093/bioinformatics/17.8.686

    CAS  Article  PubMed  Google Scholar 

  37. 37.

    Hammer J, Sturniolo T, Sinigaglia F: HLA class II peptide binding specificity and autoimmunity. Adv Immunol 1997, 66: 67–100.

    CAS  Article  PubMed  Google Scholar 

  38. 38.

    Hammer J, Bono E, Gallazzi F, Belunis C, Nagy Z, Sinigaglia F: Precise prediction of major histocompatibility complex class II-peptide interaction based on peptide side chain scanning. J Exp Med 1994, 180(6):2353–2358. 10.1084/jem.180.6.2353

    CAS  Article  PubMed  Google Scholar 

  39. 39.

    Xiao YS M.: Prediction of Genomewide Conserved Epitope Profiles of HIV-1: Classifier Choice and Peptide Representation. Statistical Applications in Genetics and Molecular Biology 2005., 4(1):

    Google Scholar 

  40. 40.

    Hattotuwagama CK, Toseland CP, Guan P, Taylor DJ, Hemsley SL, Doytchinova IA, Flower DR: Toward prediction of class II mouse major histocompatibility complex peptide binding affinity: in silico bioinformatic evaluation using partial least squares, a robust multivariate statistical technique. J Chem Inf Model 2006, 46(3):1491–1502.

    CAS  Article  PubMed  Google Scholar 

  41. 41.

    Wauben MH, van der Kraan M, Grosfeld-Stulemeyer MC, Joosten I: Definition of an extended MHC class II-peptide binding motif for the autoimmune disease-associated Lewis rat RT1.BL molecule. Int Immunol 1997, 9(2):281–290. 10.1093/intimm/9.2.281

    CAS  Article  PubMed  Google Scholar 

  42. 42.

    Muller T, Vingron M: Modeling amino acid replacement. J Comput Biol 2000, 7(6):761–776. 10.1089/10665270050514918

    CAS  Article  PubMed  Google Scholar 

  43. 43.

    Muller T, Spang R, Vingron M: Estimating amino acid substitution models: a comparison of Dayhoff's estimator, the resolvent approach and a maximum likelihood method. Mol Biol Evol 2002, 19(1):8–13.

    CAS  Article  PubMed  Google Scholar 

  44. 44.

    Liu W, Meng X, Xu Q, Flower DR, Li T: Quantitative prediction of mouse class I MHC peptide binding affinity using support vector machine regression (SVR) models. BMC Bioinformatics 2006, 7: 182. 10.1186/1471-2105-7-182

    PubMed Central  Article  PubMed  Google Scholar 

  45. 45.

    Zavala-Ruiz Z, Strug I, Walker BD, Norris PJ, Stern LJ: A hairpin turn in a class II MHC-bound peptide orients residues outside the binding groove for T cell recognition. Proc Natl Acad Sci U S A 2004, 101(36):13279–13284. 10.1073/pnas.0403371101

    PubMed Central  CAS  Article  PubMed  Google Scholar 

  46. 46.

    Smith TF, Waterman MS: Identification of common molecular subsequences. J Mol Biol 1981, 147(1):195–197. 10.1016/0022-2836(81)90087-5

    CAS  Article  PubMed  Google Scholar 

  47. 47.

    Altschul SF, Koonin EV: Iterated profile searches with PSI-BLAST--a tool for discovery in protein databases. Trends Biochem Sci 1998, 23(11):444–447. 10.1016/S0968-0004(98)01298-5

    CAS  Article  PubMed  Google Scholar 

  48. 48.

    Haussler D: Convolution Kernels on Discrete Structures. Technical Report UCS-CRL-99–10 1999.

    Google Scholar 

  49. 49.

    Mathworks: MATLAB.[]

  50. 50.

    Weston J, Elisseeff A, Bakir G, Sinz F: SPIDER: object-orientated machine learning library, v. 1.6.[]

  51. 51.

    Bengio Y, Grandvalet Y: No Unbiased Estimator of the Variance of K-Fold Cross-Validation. Journal of Machine Learning Research 2003., 2003:

    Google Scholar 

  52. 52.

    Blake JD, Cohen FE: Pairwise sequence alignment below the twilight zone. J Mol Biol 2001, 307(2):721–735. 10.1006/jmbi.2001.4495

    CAS  Article  PubMed  Google Scholar 

  53. 53.

    Rammensee H, Bachmann J, Emmerich NP, Bachor OA, Stevanovic S: SYFPEITHI: database for MHC ligands and peptide motifs. Immunogenetics 1999, 50(3–4):213–219. 10.1007/s002510050595

    CAS  Article  PubMed  Google Scholar 

Download references


DRF should like to thank the Jenner Institute, University of Oxford for the award of a Fellowship. We should like to thank Matthew N Davis, Channa K Hattotuwagama, Pingping Guan, Irini A Doytchinova, and Martin J Blythe at the Edward Jenner Institute for Vaccine Research for their assistance in the early phases of the work. Prof. M. Nielsen, Centre for Biological Sequence Analysis at the Technical University of Denmark provided the Gibbs Sampler dataset and made helpful comments. Also thanks to Heike I. Roesner, Chemistry Research Laboratories at Oxford University, for comments on the manuscript.

Author information



Corresponding author

Correspondence to Darren R Flower.

Additional information

Authors' contributions

JS conducted the data mining and modelling. DRF conceived, designed, oversaw, and interpreted the study. JS and DRF jointly drafted the manuscript. All authors read and approved the final manuscript.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Salomon, J., Flower, D.R. Predicting Class II MHC-Peptide binding: a kernel based approach using similarity scores. BMC Bioinformatics 7, 501 (2006).

Download citation


  • Major Histocompatibility Complex
  • Major Histocompatibility Complex Class
  • Kernel Method
  • Substitution Matrix
  • Major Histocompatibility Complex Allele