Volume 9 Supplement 3

Proceedings of the Second International Symposium on Languages in Biology and Medicine (LBM) 2007

Open Access

Exploiting and integrating rich features for biological literature classification

  • Hongning Wang1,
  • Minlie Huang1,
  • Shilin Ding1 and
  • Xiaoyan Zhu1Email author
Contributed equally
BMC Bioinformatics20089(Suppl 3):S4

DOI: 10.1186/1471-2105-9-S3-S4

Published: 11 April 2008



Efficient features play an important role in automated text classification, which definitely facilitates the access of large-scale data. In the bioscience field, biological structures and terminologies are described by a large number of features; domain dependent features would significantly improve the classification performance. How to effectively select and integrate different types of features to improve the biological literature classification performance is the major issue studied in this paper.


To efficiently classify the biological literatures, we propose a novel feature value schema TF*ML, features covering from lower level domain independent “string feature” to higher level domain dependent “semantic template feature”, and proper integrations among the features. Compared to our previous approaches, the performance is improved in terms of AUC and F-Score by 11.5% and 8.8% respectively, and outperforms the best performance achieved in BioCreAtIvE 2006.


Different types of features possess different discriminative capabilities in literature classification; proper integration of domain independent and dependent features would significantly improve the performance and overcome the over-fitting on data distribution.


In the general text classification, effective feature is essential to make the learning task more efficient and accurate. No degree of classifiers can make up for a lack of predictive information in the input features [1]. In bioscientific literature, where biological structures and terminologies are described in a large number of features, the situation is more serious: well-chosen features could improve the classification accuracy substantially and decrease the risk of over-fitting [2].

In the early days of biological literature classification study, most of the researchers depended on the domain experts to pick out the informative features. Regev et al. used expert-defined rules to extract features from the semi-structure text and figure legends. Besides, they utilized external lexical resources and semantic constraints to achieve a better coverage and accuracy [3]. Min Shi et al. employed two types of keywords as feature: one type was from the given evidences and the other type was manually extracted from the training texts by domain experts [4]. Moustafa M. Ghanem et al. utilized expert-edited regular expressions to capture frequently occurring keyword combinations (or motifs) within short segments of the text in a document [5]. All these approaches require the involvement of domain experts in identifying the specific textual objects and the informative templates, so that they cannot easily be automatically extended to an efficient and scale-free model on other biological datasets [6].

Recent years, fully automatic and scalable text classification algorithm provides an alternative to the previous methods. Wilbur employed unigram, bigram and all of the MeSH terms as the set of feature to represent the documents [7]. Dobrokhotv et al. utilized the words processed by the XEROX natural language processing tool as discriminating attributes [8]. Aaron et al. used “Bag of Words” model: content was tokenized and stemmed into unigram feature and modelled the samples as binary feature vectors [9].

Although all of these features catch some aspects of biological and statistical meanings, they still cannot well and automatically exploit the domain dependent information from the complex biological literature. It becomes a challenge in biological text mining field to automatically introduce higher level domain dependent features into the classification process and integrate with the lower level domain independent features.

In this paper, we investigate the issue of biological literature classification from the perspective of feature selection and integration, which is evaluated by BioCreAtIvE [10], an international evaluation in biological text mining. In IAS (Protein Interaction Article Sub-task) of BioCreAtIvE 2006, participants were asked to classify a given set of MEDLINE titles and abstracts, according to whether a document contains at least one physical PPI (Protein Protein Interaction) or not. This procedure would be extremely useful for facilitating the efficiency of manual curation since it will largely filter out the irrelevant documents. In the evaluation, one of our implemented classifiers achieved outstanding results: the Accuracy ranked at the 1st place, AUC and F-Score ranked at the 2nd place respectively.

Although the result is encouraging, the performance has dropped significantly from the 5-fold cross validation on the training set to the evaluation on the official testing set (15.2% lower by AUC, 11.8% lower by F-Score). Main differences between these two data sets are: 1) the testing documents are mainly published in 2006 while the training documents distributes evenly over the past years; 2) the relevant/irrelevant document rate in the training set is nearly 2:1 while in the testing set it is 1:1. To statistically analyze the phenomenon, we use the variance of Kullback Leibler divergence to estimate the distribution of the top 50 employed features on the training and testing sets as follows:
K L ` ( P , Q ) = 1 2 x ( P ( x ) log P ( x ) Q ( x ) + Q ( x ) log Q ( x ) P ( x ) ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXafv3ySLgzGmvETj2BSbqeeuuDJXwAKbsr4rNCHbGeaGqipu0Je9sqqrpepC0xbbL8F4rqqrFfpeea0xe9Lq=Jc9vqaqpepm0xbba9pwe9Q8fs0=yqaqpepae9pg0FirpepeKkFr0xfr=xfr=xb9adbaqaaeGaciGaaiaabeqaaeaabaWaaaGcbaGaem4saSKaemitaWKaeiiyaa2aaeWaaeaacqWGqbaucqGGSaalcqWGrbquaiaawIcacaGLPaaacqGH9aqpdaWcaaqaaiabigdaXaqaaiabikdaYaaadaaeqaqaamaabmaabaGaemiuaa1aaeWaaeaacqWG4baEaiaawIcacaGLPaaacyGGSbaBcqGGVbWBcqGGNbWzdaWcaaqaaiabdcfaqnaabmaabaGaemiEaGhacaGLOaGaayzkaaaabaGaemyuae1aaeWaaeaacqWG4baEaiaawIcacaGLPaaaaaGaey4kaSIaemyuae1aaeWaaeaacqWG4baEaiaawIcacaGLPaaacyGGSbaBcqGGVbWBcqGGNbWzdaWcaaqaaiabdgfarnaabmaabaGaemiEaGhacaGLOaGaayzkaaaabaGaemiuaa1aaeWaaeaacqWG4baEaiaawIcacaGLPaaaaaaacaGLOaGaayzkaaaaleaacqWG4baEaeqaniabggHiLdaaaa@6266@

where x is the word and phrase features employed in IAS, P(x) and Q(x) are the probability of x in the training and testing set respectively.

The result (see Table 1) demonstrates that there is great divergence between the probability distribution of features in the irrelevant document set. And only one thirds of the top 300 features selected from the training set accordingly occur in the testing set (see Figure 1). It is clear that our previously selected features are limited and sensitive to the data distribution. How to efficiently exploit the domain independent and dependent features in the biological literature and avoid the over-dependence on data distribution motivates us to have an in-depth investigation in this paper.
Table 1

KL Divergence on Training, Cross Validation and Testing Set

Unigram Feature

Relevant Probability

Irrelevant Probability

Training Set Vs Cross Validation Set



Training Set Vs Testing Set



(Top 50 features according to Chi-Square statistics)

Figure 1

Overlap of Features between Training and Testing Set

(Top 300 selected distinct features from the training and testing set according to Chi-Square statistics respectively)

The rest of the paper is organized as follows. We will introduce the detailed description of methodologies proposed in this paper in Methods section. In Results and discussion section, we will present the experiment results and analysis. In Conclusion section, we will summarize our contributions in this paper.


In this paper, we are engaged to investigate the issue from the perspective of feature selection and integration. The main contribution in this paper lies in that: we propose 1). domain independent feature value schema TF*ML and length-fixed string feature 2). domain dependent “semantic template” feature 3). efficient integrations among the features. These methods are described respectively in the following.

Probabilistic schema

The traditional TF*IDF schema [11] just takes into consideration the occurrences of words in the whole corpus, while discarding the distribution of words in different categories. Differently, we propose a novel probabilistic feature value schema TF*ML (production of Term Frequency by Maximum Likelihood) to substitute the traditional TF*IDF as follows:
T F ( t ) × M L ( t ) = T F ( t ) × ( log P ( t | c + ) log P ( t | c ) ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXafv3ySLgzGmvETj2BSbqeeuuDJXwAKbsr4rNCHbGeaGqipu0Je9sqqrpepC0xbbL8F4rqqrFfpeea0xe9Lq=Jc9vqaqpepm0xbba9pwe9Q8fs0=yqaqpepae9pg0FirpepeKkFr0xfr=xfr=xb9adbaqaaeGaciGaaiaabeqaaeaabaWaaaGcbaqcLbxacqWGubavcqWGgbGrcaaMe8UcdaqadaqaaKqzWfGaemiDaqhakiaawIcacaGLPaaajugCbiaaysW7cqGHxdaTcaaMe8Uaemyta0KaemitaWKcdaqadaqaaKqzWfGaemiDaqhakiaawIcacaGLPaaajugCbiaaysW7cqGH9aqpcaaMe8UaemivaqLaemOrayKaaGjbVRWaaeWaaeaajugCbiabdsha0bGccaGLOaGaayzkaaqcLbxacaaMe8Uaey41aqRaaGjbVRWaaeWaaeaajugCbiGbcYgaSjabc+gaVjabcEgaNjaaysW7cqWGqbaukmaabmaabaqcLbxacqWG0baDcqGG8baFcqWGJbWykmaaCaaaleqabaGaey4kaScaaaGccaGLOaGaayzkaaqcLbxacaaMe8UaeyOeI0IaaGjbVlGbcYgaSjabc+gaVjabcEgaNjaaysW7cqWGqbaukmaabmaabaqcLbxacqWG0baDcqGG8baFcqWGJbWykmaaCaaaleqabaGaeyOeI0caaaGccaGLOaGaayzkaaaacaGLOaGaayzkaaaaaa@7C62@

where t means the selected feature word, c+ and c mean the relevant and irrelevant category, P(t|c + ) and P(t|c ) mean the probability that t occurs in category c+and c respectively.

The sign of ML indicates the category relevance of the feature and the magnitude reflects the classification confidence. Following the same idea as TF*IDF to express the specificities of features in different documents, we also multiply TF by ML.

Here we do not depend on the posterior distribution of features to implement the prediction. To explain the reason we could rewrite the formula (2) as follows:
M L ( t ) = log P ( t | c + ) log P ( t | c ) = log P ( t , c + ) log P ( t , c ) + log P ( c + ) P ( c ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXafv3ySLgzGmvETj2BSbqeeuuDJXwAKbsr4rNCHbGeaGqipu0Je9sqqrpepC0xbbL8F4rqqrFfpeea0xe9Lq=Jc9vqaqpepm0xbba9pwe9Q8fs0=yqaqpepae9pg0FirpepeKkFr0xfr=xfr=xb9adbaqaaeGaciGaaiaabeqaaeaabaWaaaGcbaqcLbxacqWGnbqtcqWGmbatkmaabmaabaqcLbxacqWG0baDaOGaayjkaiaawMcaaKqzWfGaaGjbVlabg2da9iaaysW7cyGGSbaBcqGGVbWBcqGGNbWzcaaMe8UaemiuaaLcdaqadaqaaKqzWfGaemiDaqNaeiiFaWNaem4yamMcdaahaaWcbeqaaiabgUcaRaaaaOGaayjkaiaawMcaaKqzWfGaaGjbVlabgkHiTiaaysW7cyGGSbaBcqGGVbWBcqGGNbWzcaaMe8UaemiuaaLcdaqadaqaaKqzWfGaemiDaqNaeiiFaWNaem4yamMcdaahaaWcbeqaaiabgkHiTaaaaOGaayjkaiaawMcaaKqzWfGaaGjbVlabg2da9iaaysW7cyGGSbaBcqGGVbWBcqGGNbWzcaaMe8UaemiuaaLcdaqadaqaaKqzWfGaemiDaqNaeiilaWIaem4yamMcdaahaaWcbeqaaiabgUcaRaaaaOGaayjkaiaawMcaaKqzWfGaaGjbVlabgkHiTiaaysW7cyGGSbaBcqGGVbWBcqGGNbWzcaaMe8UaemiuaaLcdaqadaqaaKqzWfGaemiDaqNaeiilaWIaem4yamMcdaahaaWcbeqaaiabgkHiTaaaaOGaayjkaiaawMcaaKqzWfGaaGjbVlabgUcaRiaaysW7cyGGSbaBcqGGVbWBcqGGNbWzcaaMe8UcdaWcaaqaaKqzWfGaemiuaaLcdaqadaqaaKqzWfGaem4yamMcdaahaaWcbeqaaiabgUcaRaaaaOGaayjkaiaawMcaaaqaaKqzWfGaemiuaaLcdaqadaqaaKqzWfGaem4yamMcdaahaaWcbeqaaiabgkHiTaaaaOGaayjkaiaawMcaaaaaaaa@9CD5@
And the posterior distribution is:
Posterior ( t ) = log P ( c + | t ) log P ( c | t ) = log P ( t , c + ) log P ( t , c ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXafv3ySLgzGmvETj2BSbqeeuuDJXwAKbsr4rNCHbGeaGqipu0Je9sqqrpepC0xbbL8F4rqqrFfpeea0xe9Lq=Jc9vqaqpepm0xbba9pwe9Q8fs0=yqaqpepae9pg0FirpepeKkFr0xfr=xfr=xb9adbaqaaeGaciGaaiaabeqaaeaabaWaaaGcbaqcLbxacqWGqbaucqWGVbWBcqWGZbWCcqWG0baDcqWGLbqzcqWGYbGCcqWGPbqAcqWGVbWBcqWGYbGCkmaabmaabaqcLbxacqWG0baDaOGaayjkaiaawMcaaKqzWfGaaGjbVlabg2da9iaaysW7cyGGSbaBcqGGVbWBcqGGNbWzcaaMe8UaemiuaaLcdaqadaqaaKqzWfGaem4yamMcdaahaaWcbeqaaiabgUcaRaaajugCbiabcYha8jabdsha0bGccaGLOaGaayzkaaqcLbxacaaMe8UaeyOeI0IaaGjbVlGbcYgaSjabc+gaVjabcEgaNjaaysW7cqWGqbaukmaabmaabaqcLbxacqWGJbWykmaaCaaaleqabaGaeyOeI0caaKqzWfGaeiiFaWNaemiDaqhakiaawIcacaGLPaaajugCbiaaysW7cqGH9aqpcaaMe8UagiiBaWMaei4Ba8Maei4zaCMaaGjbVlabdcfaqPWaaeWaaeaajugCbiabdsha0jabcYcaSiabdogaJPWaaWbaaSqabeaacqGHRaWkaaaakiaawIcacaGLPaaajugCbiaaysW7cqGHsislcaaMe8UagiiBaWMaei4Ba8Maei4zaCMaaGjbVlabdcfaqPWaaeWaaeaajugCbiabdsha0jabcYcaSiabdogaJPWaaWbaaSqabeaacqGHsislaaaakiaawIcacaGLPaaaaaa@902B@

In the ML schema, the relevant/irrelevant document rate P ( c + ) P ( c ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXafv3ySLgzGmvETj2BSbqeeuuDJXwAKbsr4rNCHbGeaGqipu0Je9sqqrpepC0xbbL8F4rqqrFfpeea0xe9Lq=Jc9vqaqpepm0xbba9pwe9Q8fs0=yqaqpepae9pg0FirpepeKkFr0xfr=xfr=xb9adbaqaaeGaciGaaiaabeqaaeaabaWaaaGcbaWaaSaaaeaajugCbiabdcfaqPWaaeWaaeaajugCbiabdogaJPWaaWbaaSqabeaacqGHRaWkaaaakiaawIcacaGLPaaaaeaajugCbiabdcfaqPWaaeWaaeaajugCbiabdogaJPWaaWbaaSqabeaacqGHsislaaaakiaawIcacaGLPaaaaaaaaa@3DF1@ has been taken into consideration as a compensate factor. But in the posterior probability schema, the impact of relevant/irrelevant document rate is eliminated, according to the independent and identical distribution hypothesis, while it is not tenable in our situation (since the different relevant/irrelevant rate between training and testing set).

The essence of TF*ML schema is to fully utilize the category relevant information from the annotated samples, which cannot be inferred from the TF*IDF schema. Experiment results (see Table 2) demonstrate that ML steadily improves the discriminative capability of features.
Table 2

TF*ML Feature Value Schema.

The Precision/Recall/F-Score demonstrate classification capability of the model, and AUC (area under receiving operator characteristic curve) is to evaluate ranking capability of the model.

Feature value















(Performance under unigram feature)

String feature

In many text classification applications, it is appealing to take every document as a string of characters rather than a bag of words [12]. Especially in bioscience, the tokenizing and stemming procedure would incur undesired loss of the informative attributions, since many of the semantically related biomedical terms that share the same stem or morpheme are often not reducible to the same stems [6]. Therefore, we propose to directly utilize the length-fixed strings as feature to exploit most of the informative segments.

To the best of our knowledge, no one explicitly takes length-fixed strings as feature because of the explosion and sparse of feature space. However, the statistical analysis based on formula (1) demonstrates that distributional divergence between the training and testing set becomes much smaller under the length-fixed string feature (see Table 3). So we turn to take the fixed-length strings as feature: the length-fixed strings are extracted from the whole sequential text without considering the sentence boundaries and strictly consist of 26 lowercase English letters (all the letters are converted to the lowercase first), 10 numbers (0-9) and a white space. Chi-Square statistics [13] is employed to select out the significant features and TF*IDF is computed to build the feature vector (we substitute TF*ML for TF*IDF for further improvement).
Table 3

KL Divergence on Training, Cross Validation and Testing Set

String Feature(p=7)

Relevant Probability

Irrelevant Probability

Training Set Vs Cross Validation Set



Training Set Vs Testing Set



(Top 50 features according to Chi-Square statistics)

Table 4 lists the top 10 distinct features from the selected unigram features and string features respectively. It is apparent that the length-fixed string feature has at least the following potential advantages. First, inter-word features (e.g. phrasal effects) can be exploited automatically. The segmentation process spans the boundary of adjacent words, which could exploit information from the adjacent words. Second, intra-word features (e.g. morphological variants) could be captured. For example, string “interac” would occur in the word “interact” and “interaction”, both of which are important indicator of PPI relations. Third, the special meaning of length-fixed string feature in bio-literatures is that it exploits the slight but informative commonality from the structure of the words. For example, different terminologies in bio-literatures often share the same conflation (e.g. ‘phosph’ indicates the protein phosphorylation) and most of the suffix is informative (e.g. ‘ase’ is a common suffix to proteins that function as enzyme). The specific information is not recoverable when the general tokenizing and stemming procedure is applied.
Table 4

Top 10 Unigram Features and String Features ‘_’ means a white space

Unigram Feature

String Feature





















(According to Chi-Square statistics)

Named entities and semantic template features

Both of the above proposed methods are domain independent, which are endowed with well generalization capacity and are not necessarily limited to the bioscience domain. But introducing domain dependent features could greatly filter out the false positive samples and further improve the performance [14]. In biological literatures, named entities (words and phrases belonging to certain predefined classes, e.g. protein and gene), such as CDC42 (protein), and semantic templates (co-occurrences of a pre-specified type of relationship between entities of given types), such as “ProteinA interact with ProteinB”, are the most meaningful concepts in PPI documents and well conserve the syntactic and semantic structures in describing the protein interactions. So we introduce the named entities and semantic templates as feature to exploit the domain dependent information.

With the help of ABNER [15], a named entity recognition tool, 5 types of named entities in a given document could be identified: protein, DNA, RNA, cell types and cell line. Since the recognized entity space is large and sparse, we only utilize their types as feature to decrease the dimension of feature space without losing the universality.

After recognizing the named entities, semantic templates are ready to be extracted from the documents. We propose a novel template extraction algorithm named KeyBT, i.e. Key word B ased T emplate extraction algorithm, to extract the semantic templates describing the interaction patterns among all of the recognized entities.

Compared to the traditional local alignment algorithm, KeyBT operates differently: first locate statistical significant words as seeds, and expand the seeds in the contextual environment iteratively, finally preserve the most “powerful” templates as the result.

The flow chart of the KeyBT algorithm is as follows:
  1. 1)

    Locate the occurrences of predefined candidate keywords in each sentence; discard the sentences without any keywords; get the initial candidate sentence set S0;

  2. 2)

    Locate each entity type in S0; discard the sentences without any entities; get the initial candidate template set T0;

  3. 3)

    Iteratively normalize each template in T0: removing the redundant templates by syntax parsing; get the raw templates set T1;

  4. 4)

    Evaluating the templates in T1, filter out the templates of low quality, get the final template set Tf.


KeyBT not only depends on Chi-Square statistics to select the most distinct keywords but also utilizes ML to determine the category relevance of the keywords, because Chi-Square does not distinguish the association between features and different categories: a few high quality features of irrelevant category might be overwhelmed in the large amount of features of relevant category. Chi-Square is employed to select a raw candidate keyword list (with low threshold), and then top 50 features from both categories are preserved according to ML respectively.

We use the following formula [16] to evaluate the relevance of templates based on the balance between generality and specificity of the templates.
S ( t ) = ( β + log 2 t . pos + 0.5 t . neg + 0.5 ) × ln ( t . pos + t . neg + 1 ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXafv3ySLgzGmvETj2BSbqeeuuDJXwAKbsr4rNCHbGeaGqipu0Je9sqqrpepC0xbbL8F4rqqrFfpeea0xe9Lq=Jc9vqaqpepm0xbba9pwe9Q8fs0=yqaqpepae9pg0FirpepeKkFr0xfr=xfr=xb9adbaqaaeGaciGaaiaabeqaaeaabaWaaaGcbaqcLbxacqWGtbWukmaabmaabaqcLbxacqWG0baDaOGaayjkaiaawMcaaKqzWfGaaGjbVlabg2da9iaaysW7kmaabmaabaqcLbxacqaHYoGycaaMe8Uaey4kaSIaaGjbVlGbcYgaSjabc+gaVjabcEgaNPWaaSbaaSqaaiabikdaYaqabaqcLbxacaaMe8UcdaWcaaqaaKqzWfGaemiDaqNaeiOla4IaemiCaaNaem4Ba8Maem4CamNaaGjbVlabgUcaRiaaysW7cqaIWaamcqGGUaGlcqaI1aqnaOqaaKqzWfGaemiDaqNaeiOla4IaemOBa4MaemyzauMaem4zaCMaaGjbVlabgUcaRiaaysW7cqaIWaamcqGGUaGlcqaI1aqnaaaakiaawIcacaGLPaaajugCbiaaysW7cqGHxdaTcaaMe8UagiiBaWMaeiOBa4McdaqadaqaaKqzWfGaemiDaqNaeiOla4IaemiCaaNaem4Ba8Maem4CamNaaGjbVlabgUcaRiabbccaGiabdsha0jabc6caUiabd6gaUjabdwgaLjabdEgaNjaaysW7cqGHRaWkcaaMe8UaeGymaedakiaawIcacaGLPaaaaaa@8795@

where t.pos and t.neg are the positive/negative matching count of template t in the training set, and β is the parameter tuning the positive/negative matching rate.

When we get the final templates set Tf, we do not simply depend on the positive/negative matching rate of each template to make the prediction. Instead, we use them to build feature vectors and train a classifier.

Top 5 KeyBT-extracted templates are illustrated in Table 5.
Table 5

KeyBT-extracted Templates.

<PTN>, <DNA>, <CEL> mean protein, DNA and cell-line, E* means any words occurrence

KeyBT Templates

<PTN> E* <DNA> E* association E* <PTN>

<PTN> E* bind E* <DNA>

<PTN> E* interact E* <PTN>

<PTN> E* colocalize E* <CEL>

<PTN> E* contact E* <DNA> E* <PTN>

Compared with the local alignment algorithm that depends on the post evaluation to remove meaningless and noisy templates, the potential advantages of KeyBT algorithm are as follows: 1) KeyBT utilizes the statistical characteristic of the candidate keywords to largely remove noise before extraction; 2) KeyBT templates need not to fix the entities' type beforehand, so that it could catch the distribution of templates in both categories to discriminate both of the relevant and irrelevant categories; 3) the heuristic rules applied on the relation of named entities and candidate words (such as their sequence, the average template length and type of distinct entities) would guarantee the biological meaning of the extracted templates.

Feature integration

Experiment results of the overlap among the misclassified samples by different features show that there is great complement among different features: in many cases, the false prediction caused by one feature would be treated correctly by another one. And a single type of feature is easy to lead the classifier over-fitting on the data distribution (see Table 1 and Figure 1). Thus, the integration among different features would be beneficial. In this sense, we propose two kinds of integration from different levels: feature-level and classifier-level to integrate all of above proposed features.

We perform the feature-level integration in a typical way: normalizing each part of features and unifying them into a new feature vector. We do the normalization as follows [17]:
norm _ value = unnorm _ value min _ value  max _ value min _ value MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXafv3ySLgzGmvETj2BSbqeeuuDJXwAKbsr4rNCHbGeaGqipu0Je9sqqrpepC0xbbL8F4rqqrFfpeea0xe9Lq=Jc9vqaqpepm0xbba9pwe9Q8fs0=yqaqpepae9pg0FirpepeKkFr0xfr=xfr=xb9adbaqaaeGaciGaaiaabeqaaeaabaWaaaGcbaqcLbxacqWGUbGBcqWGVbWBcqWGYbGCcqWGTbqBcqGGFbWxcqWG2bGDcqWGHbqycqWGSbaBcqWG1bqDcqWGLbqzcaaMe8Uaeyypa0JaaGjbVRWaaSaaaeaajugCbiabdwha1jabd6gaUjabd6gaUjabd+gaVjabdkhaYjabd2gaTjabc+faFjabdAha2jabdggaHjabdYgaSjabdwha1jabdwgaLjaaysW7cqGHsislieGacqWFTbqBcqWFPbqAcqWFUbGBcqGGFbWxcqWG2bGDcqWGHbqycqWGSbaBcqWG1bqDcqWGLbqzcqqGGaaiaOqaaKqzWfGae8xBa0Mae8xyaeMae8hEaGNaei4xa8LaemODayNaemyyaeMaemiBaWMaemyDauNaemyzauMaeyOeI0Iae8xBa0Mae8xAaKMae8NBa4Maei4xa8LaemODayNaemyyaeMaemiBaWMaemyDauNaemyzaugaaaaa@7E52@

where max_value and min_value are the maximum and minimum values that are actually seen in the input feature set.

But there is an obvious defection in the above method: some lower dimensional features might be overwhelmed by the higher dimensional features (e.g. named entity feature has only 5 dimensions while length-fixed string feature has more than 10 thousand dimensions). Based on this consideration, we turn to perform the integration on the classifier level and propose two different ways to implement the integration. The first one is to integrate the output of each classifier: after training classifiers on different types of features respectively, we normalize and unify the output of each classifier into feature vectors and train a classifier. The other one is Adaboost[18], a general classifier integration method, which has two major advantages: firstly, Adaboost tunes the weight of each classifier according to its performance in each kind of training samples, which could fully utilize the discriminative capability of features; secondly, soft margin of Adaboost avoids the risk of over-fitting in the training process. These approaches well overcome the defection mentioned above.

Results and discussion

The benchmark corpus is provided by BioCreAtIvE 2006. The training set contains 3536 relevant documents (title and abstract) and 1959 irrelevant. The testing set contains 750 documents, 375 of which are labelled as relevant. All of the proposed features and integration methods are implemented on the linear-kernel SVM.

Probabilistic schema

In Table 2, TF*ML schema improves recall performance by 6.9% without losing precision compared to the traditional TF*IDF schema. The improvement validates the effectivity of exploiting the category relevance information of features and testifies ML to be a more effective and general feature value schema in general text classification applications.

String feature

In our experiment, the best performance is achieved when the string length p is set to 7. In Table 6, the length-fixed string feature (p=7) gains encouraging recall improvement by more than 12.0% compared to unigram and bigram feature. But the precision has dropped about 7.2% as the expense, which can be further compensated by employing TF*ML as feature value. The practical efficiency confirms our statistical analysis of the distribution of features and gives us insight in the selection of lower level features.
Table 6

Length-fixed String Feature (TF*IDF)






Unigram + Bigram





String (p=7)





(Performance under TF*IDF schema)

Named entities and semantic template features

In Table 7, only depending on a simple criterion that if a document contains at least one protein entity, the document should be judged relevant otherwise irrelevant, we could achieve a very high recall (0.96) with an acceptable precision (0.58). Our proposed template extraction algorithm KeyBT well captures the complex association between the keywords and named entities and achieves promising performance in term of precision (by 11.8%) and a better improvement comparing to our former approach ONBIRES templates [19], which is based on local alignment algorithm.
Table 7

Named Entity and Semantic Template Feature






Unigram + Bigram (TF*IDF)





Protein Entity occurrence





ONBIRES template





KeyBT template





Feature integration

In Table 8, feature-level integration contributes the improvement in terms of F-score by 5.2% and AUC by 5.9%; in table 9, integration based on the output of classifier achieves better improvement in terms of F-score by 5.3% and AUC by 6.3%. The best performance is reached by AdaBoost: in F-score by 11.5% and in AUC by 8.8%. Advantage of the feature integration is obvious: different types of features are independently selected from the corpus, which focus on the different aspects of feature space and reinforce each other. From the result, it is apparent that different integration methods well leverage the capability of different types of features and achieve promising improvement.
Table 8

Feature-level Integration











String + Entity





String + Template





String + Entity + Template





(Normalize each part of the features and unify them into new feature vectors)

Table 9

Classifier-level Integration.

Integration on length-fixed string feature, entity feature and template feature






Unigram + Bigram





Output based Integration










(Normalize the output of each classifier and unify them into new feature vectors)

Statistical significance test

Since the size of the evaluation corpus is not large enough, it is necessary to perform the statistical significance test to validate the reliability of our proposed features and integration methods. Here we employ s-test to evaluate the performance of systems on the pooled decisions on the individual documents/category pairs [20].

In Table 10, we can find that the proposed feature value schema TF*ML, length-fixed string feature and semantic template feature are much better than their counterpart (p value lower then 0.05), and two different level of feature integrations significantly improve the classification performance (p value lower then 0.005).
Table 10

Statistical Significance Test (s-test).

The null hypothesis is that the performance of two methods is the same; the alternative hypothesis is that the former is better than the latter.


String Vs. Unigram+Bigram


KeyBT Template Vs. Unigram+Bigram

p value





Feature Level Integration Vs. Unigram+Bigram

Classifier Level Integration Vs. Unigram+Bigram

p value



Comparison with the state of arts

In Table 11, the mean, standard deviation and best performance from BioCreAtIvE 2006 are selected from 51 runs of 19 teams. Under our feature selection and integration procedure, the performance outperforms the previous best results (F-score improved by 8.2% and AUC improved by 2.2%).
Table 11

Mean, Standard Deviation and Best Performance from BioCreAtIvE 2006 Vs Our Final Performance.

The best performance from BioCreAtIvE 2006 is selected from 51 runs of 19 teams respectively.












BioCreAtIvE 2006

Standard Deviation






Best Reported





Final Performance







The experiment results clearly demonstrate that the lower level features are endowed with better generalization capability, but hampered by lower accuracy; higher level features contain rich domain dependent information, with better specificity but poor universality. Integration of different level of features would benefit from the different aspects of the feature space, which would reinforce the domain dependent classification and overcome the bias on the data distribution.

Main contributions of this paper are as follows:
  1. (1)

    Propose novel domain independent feature value schema TF*ML and length-fixed string feature;

  2. (2)

    Introduce domain dependent features (e.g. named entities, semantic templates) into the biological literature classification, and propose a novel template extraction algorithm KeyBT;

  3. (3)

    Investigate the feature-level and classifier-level integration methods to incorporate the information from different levels and perspectives.


Now, the proposed methods are being integrated into our online service ONBIRES [21] as a pre-processing module. In the next step, we will be engaged in the aspect of incremental learning to make our approaches portable to different datasets.




This work was supported by the Chinese Natural Science Foundation under grant No. 60572084 and 60621062, National High Technology Research and Development Program of China (863 Program) under No. 2006AA02Z321, as well as Tsinghua Basic Research Foundation under grant No. 052220205 and No. 053220002.

This article has been published as part of BMC Bioinformatics Volume 9 Supplement 3, 2008: Proceedings of the Second International Symposium on Languages in Biology and Medicine (LBM) 2007. The full contents of the supplement are available online at http://www.biomedcentral.com/1471-2105/9?issue=S3.

Authors’ Affiliations

State Key Laboratory of Intelligent Technology and Systems Tsinghua National Laboratory for Information Science and Technology Department of Computer Science and Technology, Tsinghua University


  1. Forman G: An Extensive Empirical Study of Feature Selection Metrics for Text Classification, in the Journal of Machine Learning Research. Special Issue on Variable and Feature Selection 2002.Google Scholar
  2. Saeys Y, Inza I, Larranaga P: A Review of feature selection techniques in bioinformatics. In Bioinformatics. Volume 23. Oxford University Press; 2007:2507–2517.Google Scholar
  3. Regev Y, Finkelstein-Landau M, Feldman R: Rule-based extraction of experimental evidence in the biomedical domain: The KDD Cup 2002 (task 1). ACM SIGKDD Explorations Newsletter 4(2):90–92.Google Scholar
  4. Shi M, Edwin DS, Menon R, et al.: A machine learning approach for the curation of biomedical literature-KDD Cup 2002 (task 1). ACM SIGKDD Explorations Newsletter 4(2):93–94.Google Scholar
  5. Ghanem MM, Guo Y, Lodhi H, Zhang Y: Automatic scientific text classification using local patterns: KDD Cup 2002 (task 1). ACM SIGKDD Explorations Newsletter 4(2):95–96.Google Scholar
  6. Han B, Obradovic Z, Hu ZZ, Cathy WH, Vucetic S: Substring selection for biomedical document classification. Bioinformatics 2006, 22(17):2136–2142.View ArticlePubMedGoogle Scholar
  7. Wilbur JW: Boosting Naive Bayesian Learning on a Large Subset of MEDLINE. In Proceedings of AMIA Symposium. Los Angeles, CA; 918–922.Google Scholar
  8. Dobrokhotov PB, et al.: Combing NLP and probabilistic categorization for document and term selection for Swiss-Prot medical annotation. Bioinformatics 2003, 19: 91–94.View ArticleGoogle Scholar
  9. Cohehn AM: Automatically Expanded Dictionaries with Exclusion Rules and Support Vector Machine Text Classifiers: Approaches to the BioCreAtIve 2 GN and PPI_IAS Tasks. In Proceedings of the BioCreative Workshop: 22–25 April 2007. Madrid.. Edited by: Krallinger M. Spanish National Cancer Research Centre; 2007:169–174.Google Scholar
  10. BioCreAtIvE [http://biocreative.sourceforge.net/]
  11. Salton G, McGill MJ: Introduction to Modern Information Retrieval. McGraw-Hill, Inc; 1986.Google Scholar
  12. Zhang D, Lee WS: Extracting Key-Substring-Group Features for Text Classification. In the proceeding of KDD'06. Philadelphia, Pennsylvania, USA; 2006:474–483. August 20–23View ArticleGoogle Scholar
  13. Yang Y: A Comparative Study on Feature Selection in Text Categorization. School of Computer Science, Jan O.Pedrsen Verity, Inc. Sunnyvale, CA USA Carnegie Mellon University Pittsburgh, PA, USA.Google Scholar
  14. Matheus CJ: Adding Domain Knowledge to SBL through Feature Construction. In Proceedings of the Eighth National Conference on Artificial Intelligence. Boston; 1990:803–808.Google Scholar
  15. ABNER v1.5 homepage [http://pages.cs.wisc.edu/~bsettles/abner/#performance]
  16. Ding SL, Huang ML, Zhu XY: Semi-supervised Pattern Learning for Extracting Relations from Bioscience Texts. Proceedings of the 5th Asia-Pacific Bioinformatics Conference 2007, 307–316.View ArticleGoogle Scholar
  17. Lillis D, Toolan F, Collier R, Dunnion J: ProbFuse: A Probabilistic Approach to Data Fusion. In Proceedings of SIGIR'06. August. Seattle, Washington, USA; 2006.Google Scholar
  18. Ratsch G, Onoda T., Muller KR: Soft Margins for AdaBoost. Machine Learning 2001, 42: 287–320.View ArticleGoogle Scholar
  19. Huang ML, Zhu XY, Ding SL, Yu H, Li M: ONBIRES: Ontology-based biological relation extraction system. 4th Asia-Pacific Bioinformatics Conference 2006, 327–336. FEB 13–16Google Scholar
  20. Yang Y, Liu X: A re-examination of text categorization methods. In Proceedings of SIGIR'99 August. Berkley, CA USA; 42–49.Google Scholar
  21. ONBIRES Homepage [http://spies.cs.tsinghua.edu.cn:8080]


© Wang et al.; licensee BioMed Central Ltd. 2008

This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.