An empirical analysis of training protocols for probabilistic gene finders
 William H Majoros^{1}Email author and
 Steven L Salzberg^{1}
DOI: 10.1186/147121055206
© Majoros and Salzberg; licensee BioMed Central Ltd. 2004
Received: 05 November 2004
Accepted: 21 December 2004
Published: 21 December 2004
Abstract
Background
Generalized hidden Markov models (GHMMs) appear to be approaching acceptance as a de facto standard for stateoftheart ab initio gene finding, as evidenced by the recent proliferation of GHMM implementations. While prevailing methods for modeling and parsing genes using GHMMs have been described in the literature, little attention has been paid as of yet to their proper training. The few hints available in the literature together with anecdotal observations suggest that most practitioners perform maximum likelihood parameter estimation only at the local submodel level, and then attend to the optimization of global parameter structure using some form of ad hoc manual tuning of individual parameters.
Results
We decided to investigate the utility of applying a more systematic optimization approach to the tuning of global parameter structure by implementing a global discriminative training procedure for our GHMMbased gene finder. Our results show that significant improvement in prediction accuracy can be achieved by this method.
Conclusions
We conclude that training of GHMMbased gene finders is best performed using some form of discriminative training rather than simple maximum likelihood estimation at the submodel level, and that generalized gradient ascent methods are suitable for this task. We also conclude that partitioning of training data for the twin purposes of maximum likelihood initialization and gradient ascent optimization appears to be unnecessary, but that strict segregation of test data must be enforced during final gene finder evaluation to avoid artificially inflated accuracy measurements.
Background
The number of generalized hidden Markov model (GHMM) gene finders reported in the literature has increased fairly dramatically of late [1–8], and the community is now contemplating various ways to extend this attractive framework in order to incorporate homology information, with a handful of such systems having already been built (e.g., [9–12]). GHMMs offer a number of clear advantages which would seem to explain this growth in popularity. Chief among these is the fact that the GHMM framework, being (in theory) purely probabilistic, allows for principled approaches to constructing, utilizing, and extending models for accurate prediction of gene structures.
While the decoding problem for GHMM gene finders is arguably well understood, being a relatively straightforward extension of the same problem for traditional HMMs and amenable to a Viterbilike solution (albeit a more complex one), methods for optimally training a GHMM gene finder have received scant attention in the genefinding literature to date. What information is available (e.g., [2, 4]) seems to indicate that the common practice is to optimize the submodels of the GHMM independently, without regard for the optimality of the composite model.
The training of HMMs and GHMMs has traditionally been carried out using some form of maximum likelihood estimation (MLE). BaumWelch training [13], which is an instance of the wellknown expectation maximization (EM) procedure, is itself a form of MLE [14]. In the case of GHMM gene finders, one typically applies some form of MLE to each of the submodels (states) in the GHMM so as to render training features of each type (e.g., exon, intron, donor site) maximally likely under the induced (sub)model; i.e., maximizing:
for state q and for S_{ i }a feature of length d_{ i }from the stateqspecific training set T. The submodels are then merged into a composite model (i.e., the full GHMM) by observing transition probabilities between features in the training data corresponding to each of the GHMM states.
For example, an exon state in a GHMM can be trained by collecting ngram statistics (i.e., counts of nletter substrings) from known exon sequences and normalizing these into transition probabilities for an (n1)^{th}order Markov chain [15]. Similarly, intron, intergenic, and untranslated region (UTR) states can be modeled by collecting appropriate statistics from corresponding sample features and using these to train individual contentscoring models, such as Markov chains, neural networks, decision trees, etc. Signal sensors for donor and acceptor splice sites and start and stop codons can be trained by aligning known signals of the appropriate type and counting nucleotide frequencies at each position within a fixed window around the signal; converting these counts to relative frequencies produces probability estimates for use in a weight matrix or similar type of model. Transition and duration probabilities can likewise be estimated by observing appropriate frequencies in training data. All of these estimation activities can be performed independently, resulting in a GHMM consisting of distinct subsets of maximum likelihood parameters.
Such an approach does not, however, attend to the global optimality of the GHMM as a whole. Ideally, one would like to maximize the expected accuracy of the gene finder on unseen data. A reasonable approximation to this ideal would be to maximize the average probability of the gene parses in the training set:
where the collection of model parameters making up the GHMM is denoted θ and the elements (S, φ) of the training set T comprise pairs of sequences S and their known parses φ. This argmax gives us the parameterization under which the full gene parses (rather than the sequences) in the training set will be maximally likely (on average). Decomposing each parse φ into a series of (q_{ i }, d_{ i }) pairs, for state q_{ i }and state duration (i.e., feature length) d_{ i }, we get:
where P_{ e }, P_{ t }, and P_{ d }represent the emission, transition, and duration probabilities of the GHMM, respectively. Whereas the common MLE training procedure for GHMMs as described above optimizes the individual terms in the numerator of Equation 3 independently, the argmax above calls instead for these terms to be jointly tuned so as to optimize the entire ratio in parentheses. Intuitively, one can think of this alternate formulation as attempting to account for the process in the Viterbi algorithm (during later decoding) whereby the individual submodels "compete" for nucleotides (in the sense that each nucleotide can be emitted by only one submodel in any given parse, and the Viterbi algorithm chooses the final, predicted parse based on the values of the model parameters). Our hope is that by addressing the issue of submodel competition explicitly during parameter estimation, we will thereby empower the gene finder to better discriminate at a global sequence level between the features modeled by individual submodels in the GHMM, thereby producing more accurate gene predictions.
A similar optimization problem occurs in the field of speech recognition, in which systems of interacting acoustic models and language models are employed to optimally parse an audio stream into a series of discrete words. Interestingly, the trend in that field, starting with Bahl et al. in 1986 [16], has increasingly been away from the sole use of MLE and toward an alternative approach very similar to that prescribed by Equation 2 known as global discriminative training [17–19] or conditional maximum likelihood [20]. The problem also appears in a slightly different form in the related field of statistical natural language parsing, in which it has been suggested that global methods for optimizing competing stochastic grammar models may improve the accuracy of systems at the level of wholesentence parses [21]. Maximum discrimination HMMs have already been applied successfully to problems in the realm of biological sequence analysis [22], though their use in gene finding has apparently not yet seen widespread adoption. To our knowledge, the only gene finder reported to use discriminative training is HMMgene [23], a gene finder based on a nongeneralized HMM.
In light of these considerations, it is worth contemplating the possible gains in gene finder accuracy that might be obtained through the use of some form of discriminative training applied to a GHMM – that is, training aimed more directly at optimizing the ability of the gene finder to discriminate between exons and nonexons, thereby improving the expected accuracy of the gene finder's predictions. Anecdotal evidence already suggests that investigation of such methods may indeed be fruitful, as the process of manual tuning of GHMM parameters (i.e., "tweaking") after MLE training is commonly acknowledged by those with experience training GHMMbased gene finders (including our own systems). The practice of performing such tuning on the training set, especially when done iteratively, can be viewed as a manual form of gradient ascent optimization using the percentages of correctly predicted nucleotides, exons, and whole genes as surrogates for the Σ_{(S,φ)∈T} P(φS,θ) term in Equation 2.
We therefore decided to investigate the use of a simple form of global discriminative training for genefinding. We did this by building a rudimentary gradient ascent optimizer and applying it to a subset of the model parameters for our GHMMbased gene finder, TigrScan, as described in the Methods.
Results
Maximum likelihood versus discriminative training
Results on Arabidopsis thaliana
method  train  test  nucAcc  exonF  geneSn 

GRAPE  CV  CV  95 ± 1%  82 ± 2%  49 ± 3% 
GRAPE  CV  H  93 ± 1%  80 ± 2%  44 ± 3% 
GRAPE  T  T  95%  86%  57% 
GRAPE  T  H  94%  81%  48% 
MLE  CV  CV  90 ± 1%  72 ± 2%  33 ± 4% 
MLE  T  T  91%  75%  36% 
MLE  T  H  90%  71%  33% 
Results on Aspergillus fumigatus
method  train  test  nucAcc  exonF  geneSn 

GRAPE  CV  CV  88 ± 1%  54 ± 4%  35 ± 4% 
GRAPE  CV  H  88 ± 1%  51 ± 2%  29 ± 1% 
GRAPE  T  T  92%  65%  48% 
GRAPE  T  H  87%  51%  31% 
MLE  CV  CV  81 ± 3%  27 ± 8%  16 ± 5% 
MLE  T  T  88%  42%  28% 
MLE  T  H  83%  30%  18% 
The train column indicates whether training (i.e., parameter estimation) was performed on the entire training set (T) or on separate 800gene crossvalidation partitions (CV). The test column indicates whether accuracy was measured on the full training set (T), on onefifth of the training set (CV), or on the unseen data (H). We will consider the evaluation on H to be the most reliable measure of gene finder accuracy. For any row containing a CV, we report the average of five runs, where each run used a different 800gene subset of the training data for parameter estimation.
Data partitioning and cross validation
On both genomes and at all levels (nucleotide, exon, gene), accuracy measurements obtained through crossvalidation were closer to the accuracy measured on unseen data than were the measurements taken from the full training set, as we expected. This was true both with and without gradient ascent, though when gradient ascent was applied, even the crossvalidation results were slightly inflated. The latter observation is presumably attributable to the "peeking" that was permitted (see Methods), whereby the gradient ascent procedure received feedback from the 200 evaluation genes held out from the training set, T. This suggests that estimating even small numbers of parameters (in this case 29) from the test set can artificially inflate accuracy measurements on that set.
Discussion
The results presented above provide a clear demonstration that independent maximum likelihood estimation of submodel parameters is sufficiently neglectful of global GHMM behavior as to compromise gene finder accuracy. Even such a crude method as our 29parameter gradient ascent procedure proved to be effective at significantly improving accuracy over that achievable by simple MLE training. The potential for more sophisticated global discriminative training methods to produce even greater improvements is surely worthy of investigation.
It is interesting to observe that the natural language processing and speech recognition communities, from whom HMMbased methods were originally borrowed for use in bioinformatics, have been moving toward global discriminative training methods for some time. The two most popular forms of discriminative training for speech recognition are Maximum Mutual Information (MMI) and Minimum Classification Error (MCE). Both methods can be implemented using an iterative gradient ascent/descent algorithm. Our approach is most similar in spirit to that of MCE.
In the case of "pure" (i.e., nongeneralized) HMMs, expectationmaximization (EM) update formulas have been derived for both MMI and MCE. These formulas allow model parameters to be updated in an axisoblique (rather than axisparallel) manner; i.e., multiple parameters can be adjusted simultaneously, so that the optimizer is less constrained in following the direction of steepest gradient in parameter space. This may reduce the number of steps required for convergence. Indeed, more rapid convergence (in terms of numbers of reevaluation steps) has been cited as a concrete advantage of these EMstyle formulations over more generalized gradient ascent methods [23]. However, EMstyle approaches to the discriminative training problem for HMMs have typically involved a number of simplifying assumptions and/or heuristics, thereby voiding formal assurances of optimality (e.g., [17, 24, 25, 18, 26]). Furthermore, as with more generalized gradient ascent procedures, EM often tends to find only a local optimum rather than a global one [13].
In the case of GHMMbased gene finders, the advantages of EM over a generalized gradient ascent procedure may indeed be rather slim. The very flexibility which we find attractive in GHMMs can be expected to complicate the derivation of such EMlike update formulas for arbitrary GHMMbased gene finders, likely requiring additional assumptions and approximations that would further compromise the optimality of the EM procedure. It was for this reason that we decided to employ a more generalized gradient ascent method for the present study. A rudimentary gradient ascent optimizer is simple to implement, and the use of prediction accuracy as an objective function affords great convenience in approximating Σ_{(S,φ)∈T}P(φS,θ). Although P(φS,θ) can be more directly computed using a modified Forward algorithm [23], to do so would in theory be no more efficient than running the full gene finder, since the asymptotic run times of the Forward and Viterbi algorithms for GHMMs are equivalent. Nevertheless, inasmuch as the Forward algorithm provides a more direct approximation of P(φS,θ), its use for this purpose is worthy of investigation.
There are a number of other variations and enhancements which we are at present contemplating for our discriminative trainer. One of these involves the joint training of pairs of submodels in the GHMM using a maximum discrimination criterion rather than the usual one based on maximum likelihood. Although such an approach would not in itself directly attend to the global optimality of the GHMM (indeed, we already apply such an approach to our signal sensors during our socalled "MLE" training regime, as remarked earlier), it would at least seem to offer a promising direction for improving our existing optimizer and may be feasible without increasing the computational cost beyond what is practical.
For the present, we feel confident in making the recommendation that others tasked with the training of GHMM gene finders consider applying an automated gradient ascent procedure like that described here as a more systematic alternative to manual tuning of parameters following maximum likelihood training of individual submodels. Beyond the obvious advantage of likely improving gene finder accuracy, such an automated method may offer some degree of reproducibility (notwithstanding the typically stochastic nature of such methods) and uniformity for the purposes of comparing gene finders and gene finding algorithms. In addition, we urge those practicing manual tuning on their final "test" set to consider that their reported accuracy results may well be inflated as a result of "peeking" at the test set before the final evaluation – a practice that has been criticized in the field of machine learning (eg., [27]). That significant inflation was seen in our studies as a result of tuning only 29 of the ~90,000 GHMM parameters on the 200gene "test" set suggests that the phenomenon may conceivably occur to some degree even when an automated procedure is not employed.
Finally, we would like to make note of an unfortunate consequence of discriminative training of HMMs for biological sequence analysis, namely, that while the resulting models may possess improved ability for discrimination and therefore greater utility for specific tasks such as gene prediction, their suitability as representative models of biological knowledge (especially probabilistic knowledge) may well be reduced relative to models induced with simple MLE techniques. Indeed, some authors in the field of speech recognition (e.g., [20]) have noted that more accurate discrimination can sometimes be obtained by relaxing sumtoone constraints for probability distributions, thereby permitting the gradient ascent procedure to automatically discover appropriate weightings between states or inputs. This is reminiscent of the exon "optimism" parameter which we employ and which seems to have no principled justification (and indeed, we might speculate that this extraneous parameter proved useful precisely because it enabled a primitive form of discriminative training by providing an explicit "correction factor" or weighting between submodels). Thus, despite the apparent value of discriminative training in improving gene finder accuracy, our ability to extract biological knowledge by inspecting the parameters of a gene finder trained in this way may be somewhat hindered. For the present, this does not seem to be of great practical significance, but it is a consideration worthy at least of mention.
Conclusions
We have shown that discriminative training for GHMMbased gene finders is feasible using a rudimentary gradient ascent approach, and have briefly explored the relation between this method and the EMlike techniques which have been proposed in the field of speech recognition. Our experiments show that the gradient ascent method can result in a gene finder with substantially greater prediction accuracy. It is our hope that even greater gains in accuracy will result from extension and refinement of discriminative training techniques applied to GHMMbased gene finders.
Methods
Description of the GHMM
The gene finder TigrScan [8] is a GHMMbased program similar to Genie [1] and Genscan [2, 28]. The forwardstrand model contains six signal states (donor and acceptor sites; start and stop codons; promoter; polyA signal) and eight content states (intron; intergenic; 5' and 3' UTR; initial, internal, final, and single exons). The reversestrand model mirrors that of the forward strand. Four relative frequency histograms are used to estimate the duration probabilities of the four exon types; the four noncoding states are assumed to have geometric duration distributions and are therefore each parameterized by a single value representing the mean duration. Each content state is scored using a separate fifthorder Interpolated Markov Model (IMM) [29]. TigrScan offers a number of signal sensors, including WMMs, WAMs, WWAMs, and MDD trees [28] having any of the foregoing signal sensors as leaf models; for this study we used only (nonMDD) WAMs, though the order of the Markov chains within the WAMs was allowed to vary. Putative signals scoring below a given signal threshold are ignored by TigrScan. This threshold is chosen separately for each signal sensor so as to achieve a desired sensitivity Sn (Sn = TP/(TP+FN), TP = true positive count, FN = false negative count) on a training set of true and "decoy" signals. "Boosting" of signal sensors was performed by iteratively retraining each signal sensor on sets of training features in which the lowest scoring features were duplicated so as to focus the training procedure on the most difficult examples. Boosting has been found to improve signal detection in other application areas [30]. Most transitions in the GHMM are obligatory (such as "donor site → acceptor site"); of the nonobligatory transitions, sumtoone constraints and the forward/reverse strand equivalence reduce the number which can be independently varied to just four. Transitions into exon states are modified by an exon "optimism" multiplier (similar to that described in [6]) which has been seen anecdotally to be useful in improving prediction accuracy (unpublished data).
Parameters to be optimized
The total number of parameters which need to be estimated when training TigrScan is roughly 90,000; the large bulk of these are the ngram statistics comprising the IMMs used for the content sensors. As an initial attempt at applying discriminative training to TigrScan, we selected 29 of these ~90,000 parameters to subject to gradient ascent optimization. Although this is a miniscule proportion of the available parameters, our previous experiences with handtuning our GHMM on other data sets suggested that these 29 parameters exert a disproportionately large influence on the accuracy of the gene predictions. By limiting the number of parameters to be optimized we hoped to both accelerate the training procedure and also reduce the risk of overtraining. The selected parameters were:

mean intron, intergenic, and UTR lengths (3)

transition probabilities (4)

exon optimism (1)

WAM size and relative positioning (8)

WAM order (4)

signal sensitivity (1)

number of signal boosting iterations (8)

skew and kurtosis of exon length distributions
Modifications to skew and kurtosis of exon length distributions were found during early exploration to produce no improvements; these parameters were therefore left unchanged in all further experiments. All remaining parameters were estimated using standard MLE techniques.
For those runs in which gradient ascent was disabled (see below), the following methods were used to estimate the above 29 parameters: mean intron and UTR lengths as well as transition probabilities were estimated using MLE from training data; mean intergenic length was set to a fixed value based on the known intergenic lengths in the test set; exon optimism was set to zero; remaining parameters were selected so as to minimize the misclassification rate on a set of true and "decoy" signals selected from the training set.
Objective function and optimization procedure
As an objective function for use by the gradient ascent procedure, we decided to measure the accuracy of the current parameterization by running the gene finder on a subset of the training genes. Our hope was that this accuracy measure would provide a reasonable approximation of Σ_{(S,φ)∈T} P(φS,θ) by indicating roughly how often the current model θ would cause the correct parse φ to be predicted for training sequence S. We defined the nucleotide accuracy A_{ nuc }as the percentage of nucleotides correctly classified as coding vs. noncoding; A_{ exon }was defined as an average of exon sensitivity and specificity (where a predicted exon is considered correct only if both boundary coordinates were predicted correctly); and A_{ gene }was defined as the percentage of training genes which were predicted exactly correctly. These were all rounded to integral percentages between 0 and 100%. The objective function was then defined as:
f(θ) = 100A_{ nuc }+A_{ exon }+A_{ gene }. (4)
The A_{ nuc }and A_{ exon }terms were included in an effort to smooth the function, which would otherwise have been insensitive to changes not reflected in the number of genes predicted exactly correctly – i.e., a step function. Though the A_{ nuc }term was given much greater weight for this study, additional work needs to be undertaken to determine the most suitable set of weights for our objective function.
Data and experimental design
The quality of a given parameterization θ was measured by evaluating the objective function f(θ) on a heldout subset of the training set. The training set was limited to 1000 genes, and all experiments were repeated separately on two highly divergent species, the model plant Arabidopsis thaliana and the pathogenic fungus Aspergillus fumigatus. Fivefold crossvalidation was employed, so that the entire optimization procedure was carried out five times on fourfifths of the data (800 genes) and each time evaluated on the remaining onefifth (200 genes); accuracy results reported here were obtained by averaging the five sets of accuracy numbers obtained from the crossvalidation.
Several variations of this experiment were also performed. To evaluate the utility of splitting the training set and performing MLE and gradient ascent parameter estimation on separate subsets (as described above), we also performed MLE followed by gradient ascent training on the full training set T and again evaluated the induced models on H. To assess whether gradient ascent provided any improvement in accuracy we also trained a model on T using only MLE and evaluated that model on H. Although the virtues of crossvalidation have been well explored in the context of many other applications, we decided to use the above experimental design as a convenient opportunity to verify our expectation that it would also prove useful for objective analysis of gene finder accuracy.
Declarations
Acknowledgements
This work was supported in part by NIH grants R01LM06845 and R01LM007938. Thanks to I. Korf, D. Kulp, M. Brent, J. Allen, M. Pertea, M. Pop, A. Delcher, and L. Pachter for useful discussions. I. Korf provided valuable comments on a previous version of the manuscript and suggested the use of boosting in the training of signal sensors.
Authors’ Affiliations
References
 Kulp D, Haussler D, Reese MG, Eeckman FH: A generalized hidden Markov model for the recognition of human genes in DNA. In In Proceedings of the Fourth International Conference on Intelligent Systems for Molecular Biology: 12–15 June 1996 St. Louis. Edited by: States DJ, Agarwal P, Gaasterland T, Hunter L, Smith RF. Menlo Park: American Association for Artificial Intelligence; 1996:134–142.Google Scholar
 Burge C: Identification of genes in human genomic DNA. PhD thesis. Stanford University, Mathematics Department; 1997.Google Scholar
 Salamov A, Salovyev V: Ab initio gene finding in Drosophila genome DNA. Genome Res 2000, 10: 516–522.PubMed CentralView ArticlePubMedGoogle Scholar
 Cawley SE, Wirth AI, Speed TP: Phat – a gene finding program for Plasmodium falciparum . Mol Biochem Parasitol 2001, 118: 167–174.View ArticlePubMedGoogle Scholar
 Majoros WM, Pertea M, Antonescu C, Salzberg SL: GlimmerM, Exonomy and Unveil: three ab initio eukaryotic genefinders. Nucleic Acids Res 2003, 31: 3601–3604.PubMed CentralView ArticlePubMedGoogle Scholar
 Stanke M, Waack S: Gene prediction with a hidden Markov model and a new intron submodel. Bioinformatics 2003, 19: II215II225.View ArticlePubMedGoogle Scholar
 Korf I: Gene finding in novel genomes. BMC Bioinformeltics 2004, 5: 59.View ArticleGoogle Scholar
 Majoros WM, Pertea M, Salzberg SL: TigrScan and GlimmerHMM: two opensource ab initio eukaryotic gene finders. Bioinformatics 2004, 20: 2878–2879.View ArticlePubMedGoogle Scholar
 Korf I, Flicek P, Duan D, Brent MR: Integrating genomic homology into gene structure prediction. Bioinformatics 2001, (Suppl 17):140–148.
 Yeh RF, Lim LP, Burge CB: Computational inference of homologous gene structures in the human genome. Genome Res 2001, 11: 803–816.PubMed CentralView ArticlePubMedGoogle Scholar
 Alexandersson M, Cawley S, Pachter L: SLAM: Crossspecies gene finding and alignment with a generalized pair hidden Markov model. Genome Res 2003, 13: 496–502.PubMed CentralView ArticlePubMedGoogle Scholar
 Zhang L, Pavlovic V, Cantor CR, Kasif S: Humanmouse gene identification by comparative evidence integration and evolutionary analysis. Genome Res 2003, 13: 1190–1202.PubMed CentralView ArticlePubMedGoogle Scholar
 Baum LE: An inequality and associated maximization technique in statistical estimation for probabilistic functions of Markov processes. Inequalities 1972, 3: 1–8.Google Scholar
 Rabiner LR: A tutorial on hidden Markov models and selected applications in speech recognition. Proc of the IEEE 1989, 77: 257–285.View ArticleGoogle Scholar
 Krogh A: An introduction to hidden Markov models for biological sequences. In In Computational Methods in Molecular Biology. Edited by: Salzberg SL, Searls DB, Kasif S. Amsterdam: Elsevier Science BV; 1998:45–62.View ArticleGoogle Scholar
 Bahl LR, Brown PF, de Souza PV, Mercer RL: Maximum mutual information estimation of hidden Markov model parameters for speech recognition. In In Proceedings of the International Conference on Acoustics, Speech and Signal Processing. IEEE Computer Society Press; 1986:49–52.View ArticleGoogle Scholar
 Krogh A: Hidden Markov models for labeled sequences. In Proceedings of the Twelfth IAPR International Conference on Pattern Recognition: 9–12 October 1994. Jerusalem. Piscataway: IEEE Computer Society Press; 1994:140–144.Google Scholar
 Jelinek F: Statistical Methods for Speech Recognition. Cambridge: Bradford Books; 1997.Google Scholar
 Schlüter R, Macherey W, Müller B, Ney H: Comparison of discriminative training criteria and optimization methods for speech recognition. Speech Communication 2001, 34: 287–310.View ArticleGoogle Scholar
 Johansen FT: A comparison of hybrid HMM architectures using global discriminative training. In In Proceedings of the Fourth International Conference on Spoken Language Processing: 3–4 October 1996 Philadelphia. Piscataway IEEE Computer Society Press; 1996:498–501.View ArticleGoogle Scholar
 Toutanova K, Mitchell M, Manning CD: Optimizing local probability models for statistical parsing. In In Proceedings of the Fourteenth European Conference on Machine Learning (ECML 2003). New York: Springer Verlag; 2003:409–420.Google Scholar
 Eddy S, Mitchison G, Durbin R: Maximum discrimination hidden Markov models of sequence consensus. J Comput Biol 1995, 2: 9–23.View ArticlePubMedGoogle Scholar
 Krogh A: Two methods for improving performance of an HMM and their application for gene finding. In In Proceedings of the Fifth International Conference on Intelligent Systems for Molecular Biology: 21–25 June 1997 Halkidiki, Greece. Edited by: Gaasterland T, Karp P, Karplus K, Ouzounis C, Sander C, Valencia A. Menlo Park: American Association for Artificial Intelligence; 1997:179–186.Google Scholar
 Reichl W, Ruske G: Discriminative training for continuous speech recognition. In In Proceedings of the Fourth European Conference on Speech Communication and Technology (EUROSPEECH95): 18–21 September 1995 Madrid. Amsterdam: Institute of Phonetic Sciences; 1995:537–540.Google Scholar
 Normandin Y: Maximum mutual information estimation of hidden Markov models. In In Automatic Speech and Speaker Recognition. Edited by: Lee CH, Soong FK, Paliwal KK. Norwell: Klewer Academic Publishers; 1996:58–81.Google Scholar
 Markov K, Nakagawa S, Nakamura S: Discriminative training of HMM using maximum normalized likelihood algorithm. In In Proceedings of the International Conference on Acoustics, Speech and Signal Processing: 7–11 May 2001 Salt Lake City. IEEE Computer Society Press; 2001:497–500.Google Scholar
 Salzberg SL: On comparing classifiers: a critique of current research and methods. Data Mining and Knowledge Discovery 1999, 1: 1–12.Google Scholar
 Burge C, Karlin S: Prediction of complete gene structures in human genomic DNA. J Mol Biol 1997, 268: 78–94.View ArticlePubMedGoogle Scholar
 Salzberg SL, Pertea M, Delcher AL, Gardner MJ, Tettelin H: Interpolated Markov models for eukaryotic gene finding. Genomics 1999, 59: 24–31.View ArticlePubMedGoogle Scholar
 Friedman J, Hastie T, Tibshirani R: Additive logistic regression: a statistical view of boosting. Annals of Statistics 2000, 38: 337–374.View ArticleGoogle Scholar
 Manning CD, Schütze H: Foundations of statistical natural language processing. Cambridge: MIT Press; 1999.Google Scholar
 Altschul SF, Gish W, Miller W, Myers EW, Lipman DJ: Basic local alignment search tool. J Mol Biol 1990, 215: 403–410.View ArticlePubMedGoogle Scholar
Copyright
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.