Volume 13 Supplement 11
Event extraction of bacteria biotopes: a knowledge-intensive NLP-based approach
© Ratkovic et al.; licensee BioMed Central Ltd. 2012
Published: 26 June 2012
Bacteria biotopes cover a wide range of diverse habitats including animal and plant hosts, natural, medical and industrial environments. The high volume of publications in the microbiology domain provides a rich source of up-to-date information on bacteria biotopes. This information, as found in scientific articles, is expressed in natural language and is rarely available in a structured format, such as a database. This information is of great importance for fundamental research and microbiology applications (e.g., medicine, agronomy, food, bioenergy). The automatic extraction of this information from texts will provide a great benefit to the field.
We present a new method for extracting relationships between bacteria and their locations using the Alvis framework. Recognition of bacteria and their locations was achieved using a pattern-based approach and domain lexical resources. For the detection of environment locations, we propose a new approach that combines lexical information and the syntactic-semantic analysis of corpus terms to overcome the incompleteness of lexical resources. Bacteria location relations extend over sentence borders, and we developed domain-specific rules for dealing with bacteria anaphors.
We participated in the BioNLP 2011 Bacteria Biotope (BB) task with the Alvis system. Official evaluation results show that it achieves the best performance of participating systems. New developments since then have increased the F-score by 4.1 points.
We have shown that the combination of semantic analysis and domain-adapted resources is both effective and efficient for event information extraction in the bacteria biotope domain. We plan to adapt the method to deal with a larger set of location types and a large-scale scientific article corpus to enable microbiologists to integrate and use the extracted knowledge in combination with experimental data.
Scientific documents on bacteria biotopes describe the locations from which bacteria have been isolated. The amount of knowledge about bacteria and their isolation sites is rapidly growing due to the advancement of species identification technologies. A better understanding of the interaction of bacteria with their environment, phenotype, metabolism and genetics could be achieved by studying their correlations with their biotope properties. The availability of an increasing amount of biological data makes such large-scale studies possible . A formal representation of the taxa associated with their locations is a first critical step. Descriptions of bacteria isolation sites that are available in public biology databases and biological resource centers are incomplete and rarely normalized . Scientific papers and genome sequencing project web pages are an invaluable source of such information.
The BioNLP 2011 Bacteria Biotope shared task (BB) has defined bacteria location identification as an event extraction task. The participants are required to precisely identify bacteria and their locations in the form of text bound annotations of scientific documents. There are eight location types to be predicted: Host, Host-part, Geographical, Food, Medical, Soil, Water and Environment. Each of these eight location types must be related to a bacterium by a localization relation. Host and Host-part locations must be related by a Part-of relation as well.
A lexicon-based approach appeared more appropriate to predict bacteria, host and geographical entity names because these are predominantly proper names. Additionally, these locations are more subject to morphological variations, whereas other location names display syntactic variability.
In the training corpus, we observed that not only bacteria species names but also higher level taxa (families) and lower level taxa (strains) were annotated as event arguments. For bacteria recognition, we used the prokaryote tree of the NCBI taxonomy  as the main resource because it includes both scientific and common names, and it is kept up to date. We applied rewriting rules to extend it with standard abbreviations of scientific names (e.g., Bacillus subtilis = B. subtilis). It was also manually enriched by 213 bacterium taxa from the training corpus, in particular, non-standard abbreviations (e.g., Chl. = Chlorobium, ssp. = subsp, GSB for green sulphur bacteria) and plurals (e.g., Vibrios as the plural for Vibrio). The original NCBI taxonomy contains 270,190 bacteria names while the extended version contains 407,147.
Determining the correct boundaries of the bacteria names was a major issue because the strain names found in the corpus do not always follow conventional nomenclature rules . Five specific rules were developed to account for irregular forms that could not be found by exact match. (1) Variation in nomenclature naming, such as the inclusion of the word strain or the abbreviations sp. or spp. (e.g., Bordetella petrii strain DSM12804; Borrelia spp.). The rule looks for both standard and non-standard nomenclature names for bacteria. (2) Variations in word order (e.g., LB400 of Burkholderia xenovorans instead of Burkholderia xenovorans LB400). This rule looks for text segments that contain all tokens of a given strain name in any order and that may contain function words, such as of, as in LB400 of Burkholderia xenovorans. (3) Parentheses (e.g., Tropheryma whipplei (the Twist strain) instead of Tropheryma whipplei strain Twist). This rule handles bacteria names with the strain name in parenthesis, which may contain other function words. (4) Partial strain names (strain DSMZ 245 standing for Chlorobium limicola strain DSMZ 245 T). (5) Bacteria names that contain modifiers such as antimicrobial-resistant C. coli or L. pneumophila serogroup 1. Such modifiers are not found in the NCBI lexicon. In the training and development sets, 107 instances of bacteria were extracted using the above rules. These five rules improve bacteria recall by 1.7 points (from 82.6% to 84.3%). Furthermore, they considerably improve the overall system recall by 13.7 points (from 38.4% to 52.1%).
The main source of error in bacteria name prediction is the ambiguity caused by the use of genus and strain name abbreviations within the same text. This error frequently occurs when the species scientific name is abbreviated using only the first word, which is also the genus name. For example, Bartonella henselae species is abbreviated as Bartonella. Unfortunately, the Bartonella genus is also mentioned in the same text and thus yields an ambiguity between the species anaphor and the genus name. This issue is further addressed in the Anaphora Resolution section.
The same strategy was used for host name recognition. We considered eukaryote species as bacteria hosts, neglecting bacteria as possible candidate hosts. The eukaryote subtree of the NCBI taxonomy was the main source of host names (811,392), which was complemented with standard abbreviations (2,110,434). Our system uses a list of 125 common English words to remove ambiguous taxon names, such as Indicator (honeyguides) and Dialysis (xylophages insect), because such terms are more frequent as common words than as taxon names. Removing these common words significantly increased the host precision by 16.6 points (from 38.8% to 55.4%). This list was enriched using 412 non-taxonomic host groups (e.g., herbivores), progeny names (e.g., calf) and human categories (e.g., patient). The resulting host name list contains 2,078,092 scientific names and 32,186 common names.
At first, we considered using the very rich resource GEOnet Names Server (GNS)  for geographical name recognition. However, it contains too many ambiguous names to be directly usable for short-term development, as required by the challenge. For the challenge, we used the short list of Agrovoc geographical names . The official recall was 29%, the worst among participant systems. The current version of the Alvis geographical name recognition component uses the country list of DBpedia , which has 1,201 entries and includes country and US state names. An additional contextual pattern increases the recall of the geographical name recognition, which considers the proper names preceding country or U.S. state names as geographical locations as well, for example, Mono Lake, California. The new developments slightly improved the overall recall by 2.4 points. However, there is still room for improvement, which can be achieved using more sophisticated patterns and larger resources.
Symbolic syntactic-semantic method
The other location entities of types Soil, Water, Medical, Environment and Host-Part cannot be handled by lexicon-based methods only. Available lexicons are necessary resources, but they cannot account for the vocabulary variety given that a location can be any physical matter. Moreover, as observed in the corpus, these location terms are noun phrases with adjectival and noun modifiers, verbal and prepositional complements subject to deep syntactic and semantic variations. We propose a robust location entity extraction method that overcomes both the high degree of variation and the incompleteness of the lexicons. The first step extracts all candidate terms from the corpus. The second step assigns a location type to the candidate terms and eliminates the irrelevant ones. This step exploits the morpho-syntactic structure of the terms and the location types derived from the location lexicon.
For term extraction, we used Alvis with BioYaTeA , an extended version of YaTeA  that has been adapted to the biology domain. We analyzed the training dataset to set BioYaTeA parameters and adapt it to the task. BioYaTeA parameters are a set of modifiable constraints that include boundaries, morpho-syntactic patterns and domain specific post-filters. We restricted the morpho-syntactic patterns to noun phrases with adjectival modifiers (e.g., microbial mat). Prepositional complements in the training dataset location terms are rather rare (e.g., breaks in the skin), except of prepositional complements (e.g., nodules of plants). The boundaries were set to include only the of preposition and exclude other prepositions, such as with or at, which may yield syntactic attachment errors (e.g., contaminating an open wound with sea water). We thus preferred the risk of incomplete terms to incorrect prepositional attachments. BioYaTeA produces a set of candidate terms with their syntactic structure in the form of head-modifier relations (e.g., geothermal<=M> environment<=H>).
The type assignment allocates types to BioYaTeA terms, with respect to the location lexicon, where each entry has a location type among Soil, Water, Medical, Food, Environment and Host-Part. Head identification is a crucial point for our type assignment method. We observed that, in most cases, the head of the candidate location compound conveys the location type information. However, some terms have ambiguous heads with respect to the location types (i.e., canal as Host-Part or Water) or uninformative heads, i.e., those that convey little or no type information, such as area or sample.
Examples of type assignment.
sludge of a zinc decantation tank
Term head and disambiguation
Host-Part (not Water)
marine (environment is an uninformative head)
Examples of disambiguation and location boundary adjustment rules.
IF dead, decaying in term THEN type =Env
dead animal (Host) becomes dead animal (Env)
IF nationality before term (Env) THEN include nationality
oil field (Env) becomes German oil field (Env)
IF irrelevant modifier in term THEN delete modifier
infected rodent (Host) becomes rodent (Host)
Adjustment of term boundaries
The BB task guidelines specify that the location span should exclude information that is irrelevant to bacteria living conditions, such as infected in infected tick. The method adjusts the location boundaries by removing irrelevant modifiers from the extracted terms. We manually built a list of these modifiers by examining the training set, the habitat and isolation site fields of the GOLD database .
The guidelines also specify that location spans do not overlap. However, it is possible that the location terms predicted by Alvis are nested because BioYaTeA extracts not only the complete terms, but also their subterms (i.e., constituents). To choose the right subterm, the method selects the maximum disjoint terms among the extracted location terms. For instance, the term human gastrointestinal tract has four location subterms: gastrointestinal tract, human, gastrointestinal and tract. The maximum disjoint location terms are gastrointestinal tract and human. The others are discarded because they are overlapping.
Resources for type assignment
The location lexicon is processed before location prediction to compute its term heads and types. The term head is computed by BioYaTeA. The assignment strategy of types to lexicon entries depends on the structure of the lexicon, more precisely whether it is flat or hierarchical. We experimented with three different resources: the flat list of all tagged location terms of the training and development sets (.a2 files) referred to a DTa2, the Microorganism Biotope Termino-Ontology (MBTO) and the EnvO habitat ontology . MBTO focuses on bacteria biotopes and phenotypes with a strong emphasis on the physico-chemical properties of the habitats. MBTO had previously been developed at the Institut National de la Recherche Agronomique (INRA). EnvO includes eukaryote habitats and is less focused on bacteria habitat descriptions.
DTa2 location terms of the task corpus are already annotated with types. To associate MBTO and EnvO concepts with the BB task types, we took advantage of the hierarchical structure of the ontologies. We manually associated the high level nodes of the location hierarchies to the eight location types. The types of the lower level concepts were then inferred. For instance, the concept aquatic environment is tagged as Water in the ontology, and all of its descendants, including lake, sea and ocean, are consequently of the same type. Local exceptions were manually handled. Once the ontologies have been processed, they can be used by the corpus type assignment algorithm.
Experiments with the typing method
Number of terms in the test corpus per type assignment method and per resource.
% of corpus terms
Main head of term
Total head matching
There are a broad range of methods that aim to generate semantic annotations of entities using types and ontologies. Among them, our type assignment method belongs to the class of non-statistical terminological methods that rely on partial parsing and the semantic analysis of terms . It is strongly related to the head-similarity analysis, as previously described . The way our method associates corpus terms to lexicon entries is similar to MetaMap . MetaMap tags biomedical corpora with the UMLS Metathesaurus using syntactic analysis, which takes into account lexical heads of terms for matching lexicon-corpus terms. However, as opposed to MetaMap, the aim of our lexicon-corpus term matching is not to identify the semantically closer terms but to use the matching result to infer the relevant type of the corpus terms at a coarse grain level. Therefore, a resource less extensive than UMLS is required, which is indicated by the favorable results obtained using DTa2 (Results section). This matching does not have to be as precise, and consequently, the computation of term variation, as performed by MetaMap, is not necessary.
The extraction of relations from text has led to a large amount of work involving an increasing amount of linguistic analysis, including syntactic analysis, such as dependency parsing  and discriminant classification machine learning methods, e.g., ILP , SVM [21–23], KNN with weak supervision  or maximum entropy .
For the prediction of localization and Part-of events, we used a straightforward and efficient strategy based on the co-occurrence of arguments and trigger words within a sentence, as previously described . The system predicts a localization event between a bacterium and a location when a bacteria name (or a bacteria anaphor), a location and a trigger word occur together in the same sentence. Similarly, the system predicts a Part-of event between the host and the host part when a host name, a host part term, a bacterium's name (or a bacteria anaphora) and a trigger word occur together in the same sentence. Requiring the occurrence of a bacterium and a trigger for predicting a Part-of event may appear superfluous. However, removing the bacterium's name and trigger word co-occurrence constraint negatively affects the precision of the Part-of event extraction without improving the recall. The trigger words denote the bacteria residence (e.g., inhabit, lives, niche, environment), its dissemination and contamination means (e.g., colonize, ingest), its pathogenic effects (e.g., chronic, disease, pathogen) and bacteria sampling (e.g., discover, isolate). The trigger word list was designed by automatically ranking words in the training corpus sentences that contain both a bacteria name and a location term. The ranking criterion used was the information gain with respect to whether the sentence actually contained an event. The ranked list was manually adjusted by removing words that do not designate a location relation and adding 35 domain knowledge words (e.g., outbreak, flora, epidemic). In total, 55 trigger words were collected. The list contains word stems to account for morphological variations and variations in part of speech categories among the triggers. To measure the potential negative effect of trigger words on the recall of the localization event extraction, we tested the algorithm without their use (Results section).
Anaphors are encountered frequently in the BB task corpus, especially for bacteria and to a much smaller extent for hosts. Our effort focused on bacteria anaphora resolution. The location relation extraction method, as described above, assumes that the relation arguments (location and bacterium, or anaphor of the bacterium) occur in the same sentence. From a total of 2,296 sentences in the training corpus, only 363 sentences contained both a location and an explicit bacterium, while 574 mention only a location. Thus, anaphora resolution is critical for location event extraction.
The style of some of the documents is rather relaxed, and the antecedent may be ambiguous even for a human reader. A manual examination of the training corpus indicated that the most frequent bacterial anaphors are not pronouns, but hyperonym definite expressions, either higher level taxa often preceded by a demonstrative determinant (This bacteria, This Clostridium) or sortal anaphors (genus, organism, species, strain). Both types of anaphors are commonly found in biology texts . For anaphora detection and resolution, a pattern-based approach was preferred to machine learning because the constraints for relating anaphors to antecedent candidates of the same taxonomy level were mainly semantic and domain-dependent. Additionally, the annotation of anaphors was not provided in the training corpus.
In general, the resolution of anaphors has three steps: (1) the identification of the anaphoric expression, (2) the search for the set of possible antecedents and (3) the selection of the antecedent from the set . In our case, the set of possible antecedents is computed beforehand because only bacteria anaphors are handled and the bacteria mentions are detected prior to anaphora resolution. Thus, the Alvis anaphora resolution is only required to perform the first and final steps.
We handled three types of anaphors in the corpus: standard anaphors that have a unique bacteria name as an antecedent (example 1), bi-anaphors that have two bacterial names as antecedents (example 2) and a higher taxon name being used as an anaphoric expression to refer to a lower taxon, which we named name taxon anaphors (example 3).
Example 1: Anaphor with a unique antecedent (BTID-60106)
C. coli is pathogenic in animals and humans. People usually get infected by eating poultry that contained the bacteria , eating raw food, drinking raw milk, and drinking bottle water [...].
Example 2: Anaphor with two antecedents (BTID-60106)
C. coli is usually found hand in hand with its bacteria relative, C. jejuni . These two organisms are recognized as the two most leading causes of acute inflammation of intestine in the United States and other nations.
Example 3: Name taxon anaphor (BTID-10090)
Ticks become infected with Borrelia duttonii while feeding on an infected rodent. Borrelia then multiplies rapidly, causing a generalized infection throughout the tick.
Example 4: Localization sentence without anaphor (BTID-60051)
In the 1600s anthrax was known as the "Black bane" and killed over 60,000 cows. The first step automatically identifies potential anaphors in the corpus given a list of pronouns, sortal anaphors and taxa [29, 30].
The final step selects the most likely bacteria antecedent from the set of possible candidates to relate it to the anaphor. Much of the work performed on anaphora resolution describes the different morphological features of the anaphor/antecedent pair to match them . These features include gender and number, which are useful in narrowing down the antecedent set. In our case, these features would not be useful because we are only treating bacteria anaphors and the word bacteria is often used to refer to both a single bacterium and the species. Therefore, we focus on two other features: the distance between the anaphor and its antecedent and the salience of the antecedent candidate in the sentence.
The antecedent is usually found close to the anaphor to maintain the coherence of the discourse. Therefore, our method ranks the antecedent candidates according to the anaphoric distance counted in the number of sentences. The closer the antecedent is to the anaphor, the more likely it is to be selected. If more than one bacterium is found in a given sentence, their position is a discriminant. Centering theory states that in a sentence the most prominent entities and, therefore, the most probable antecedent candidates are in the order: subject > object > other position . In English, due to the SVO order of the language, the subject is most often found at the beginning of the sentence, followed by the object and then the others. Therefore, the method retains the leftmost bacterium in the sentence when searching for the best antecedent candidate. For anaphors that require two antecedents, we use these same criteria but search for two bacteria in each sentence or paragraph instead of one. For taxon anaphora, we search for the presence of a lower taxon in the document found before the anaphor that is compatible according to the species taxonomy.
Number of anaphors found in each corpus per anaphor type.
The anaphora resolution algorithm allowed us to retrieve more sentences that contain both a bacterium and a location. Of the 574 sentences that contain only a location, 436 (76%) were found to also contain an anaphor related to at least one bacterium. The remaining 138 sentences are cases where there is no bacteria anaphor or the bacterium mention is implicit. The bacterium is frequently referred to through its action. For instance, in example 4 above, the anthrax bacterium name could be derived from the name of the disease that it causes.
Alvis system scores for event extraction on the BB task.
45 (23) +22
66 (79) -13
53 (36) +17
52 (45) +7
46 (45) +1
49 (45) +4
Alvis system scores for entity recall, recall, precision and F-score on the BB task by location type.
The overall event extraction score improves by 4 points (Table 5). This improvement is mainly due to an increase in recall while the precision, more or less, stays the same. However, for the Part-of relation, the recall significantly increases (23% to 45%) while the precision decreases (79% to 66%), resulting in a significantly higher F-score. The localization event recall (Table 6, row 2) is on average 20% lower for all types than the location entity recall (Table 6, row 1), which indicates that the events are equally difficult to predict independently of the argument types. The prediction of hosts and bacteria entities achieved a very good recall of 84 and 82 (row 1), respectively, indicating the relevance of the NCBI resource and additional patterns. Geographical locations were poorly recognized (29%), but the new combination of an appropriate DBpedia dictionary and the contextual patterns greatly improved the recall by 41 points, probably slightly decreasing the entity precision, as shown by the poor event precision of geographical entities (36%). Environment arguments remain harder to predict due to their high naming variability (61%).
The localization event precision (Table 6, row 3) is more difficult to analyze than the recall because the precision of the argument prediction is not measured. Many sources of error may be involved, such as incorrect arguments, incorrect anaphora resolution, relation to the wrong bacterium among several or the lack of a relation.
The investigation of the test corpus also shows that in some cases environmental locations were mentioned in the documents as potential sources of industrial applications without actually being bacteria isolation sites. The Alvis system does not handle such modalities. For instance, in Some enzymes of Thermus species [...] Other fields of application [...], waste treatment, pulp and paper manufacture, and animal feed and human food (BTID-50167), the Alvis system erroneously relates Thermus to waste treatment, paper manufacture, animal feed and human food.
The event prediction results denoted by the F-score vary with respect to the type of location (row 4). The best result is obtained for the Soil type (70.3%), followed by Host (59%) and Host-part (46.8%). The lowest F-score is obtained for the Environment type (30.5%), followed by Geographical (33.5%) and Water (39%), which also had the worst entity recall. These results indicate that the recognition of entities varies in difficulty. Most of the soil terms have the word soil as their head, making them easier to type. Conversely, the environments are the most unpredictable due to their diversity. Compared with the official results, the new results increased for all of the types, except for Water. This improvement is explained by the refinement of the syntactic analysis and head word identification method, which is critical for location type assignment. The Water exception deserves further investigation. Considering the low number of test examples (21), it may be due to the introduction of a few type errors as a consequence of the MBTO expansion.
Results by resource
Scores obtained by different resources and the symbolic syntactic-semantic type assignment method.
Scores obtained by different resources and the exact match type assignment method.
Experimental results per method
The symbolic syntactic-semantic algorithm allows us to type previously unobserved terms (an average recall increase of 8 points compared with exact match; see Tables 7 and 8). However, the mapping of terms to heads is not one-to-one, possibly resulting in a term having two or more types. This method, in turn, can introduce some ambiguity, resulting in a lower precision (average precision decrease of 2.5 points; see Tables 7 and 8). Overall, the typing algorithm grants a significant 4.9 % average increase.
For the baseline, the entity recall is rather high (61.8 %), indicating that using well-adapted and complete resources allows us to recover a high proportion of arguments. The symbolic syntactic-semantic typing method also significantly increases the entity recall, demonstrating that the approach we use is effective and relevant for this task. Overall, Figure 3 shows the contribution of each of the strategies presented in this paper. Each of the four modules presented displays an important (recall and/or precision) and significant improvement of the overall score.
Future work will focus on the improvement of the prediction of location entities and events using more syntactic information.
Term extraction associated with the symbolic syntactic-semantic typing method appears to be very efficient for predicting locations, including unobserved entities. The similarity of the training and test set vocabulary resulted in a good performance of the argument prediction using the DTa2 annotated terms. However, MBTO obtained similar results. With the typing method, MBTO should, therefore, demonstrate higher prediction capabilities than DTa2 on new corpora for which no manual annotations are provided.
Regardless of the lexicon richness, the typing strategy excludes all terms with unknown heads, which may be critical in cases where the corpus may not be representative of the available lexicon. In the future, we will overcome this potential limitation by studying the benefit of using linguistic markers, such as exemplification structures, for recovering additional location terms. For example, in the exemplification expression heated organic materials, such as compost heaps, rotting hay, manure piles or mushroom growth medium, our method would correctly type not only heated organic materials as Environment but also the other location examples of the enumeration despite their unknown heads.
Alvis predicts localization and Part-of events by co-occurrence of the arguments with trigger words within the same sentence. This method has a negative effect on the precision measure because some pairs are irrelevant, as in example 5. Two hosts are predicted for the Baumannia bacterium, namely the leafhopper and the plant. However, only the first one is correct.
Example 5 (BTID-10075)
Baumannia cicadellinicola. This newly discovered organism is an obligate endosymbiont of the leafhopper insect Homalodisca coagulata (Say) , also known as the Glassy-Winged Sharpshooter , which feeds on the xylem of plants.
It has been shown that the use of syntactic dependencies to extract biological events, such as protein-protein interactions, improves the results of such systems [32, 23, 20] because syntactic dependencies indicate semantic roles. In example 6, syntactic dependencies would be useful to distinguish the host among the entities by searching for the one that is directly related to the bacteria by an agent - action - target relation. The use of syntactic dependencies could offer a deeper examination of the syntax and semantics, thereby allowing for a more refined extraction of bacteria localization and host-host part relations. However, our preliminary experiments with syntactic parsing and machine learning indicate that the high variability of syntactic structures of corpus sentences may be a significant obstacle for the induction of efficient prediction rules. This idea has yet to be investigated.
The outcome of this work concerns both the biology and the information extraction domains. The promising performance of the Alvis system on the BB task shows that a combination of semantic analysis and domain-adapted resources is a good strategy for information extraction in the biotope domain. We present an argument prediction method based on term extraction and linguistic analysis of EnvO and MBTO habitat resources. The results obtained using MBTO indicate that our argument prediction method is efficient even without any training data. Its potential application scope should be far beyond the BB task and deserves further evaluation on large collections of scientific papers that are rich in biotope information. We have also shown that the anaphora resolution algorithm considerably improves the event extraction by considering not only usual anaphoric expressions, such as pronouns, but also hyperonyms and cardinality. Definite anaphoric expressions are not specific to the BB corpus and are frequent in technical and scientific documents in experimental domains. The benefit of a combination of the two methods still has to be evaluated in other information extraction domains. A further characterization of the corpus and events for which the method is adapted must also be performed.
This work was partially supported by the National Oseo Quaero project.
This article has been published as part of BMC Bioinformatics Volume 13 Supplement 11, 2012: Selected articles from BioNLP Shared Task 2011. The full contents of the supplement are available online at http://www.biomedcentral.com/bmcbioinformatics/supplements/13/S11.
- Korbel JO, Doerks T, Jensen LJ, Perez-Iratxeta C, Kaczanowski S, Hooper SD, Andrade MA, Bork P: Systematic association of genes to phenotypes by genome and literature mining. PLoS Biology. 2005, 3: e134-10.1371/journal.pbio.0030134.PubMed CentralView ArticlePubMed
- Field D, Garrity G, Selengut J, Sterk P, Tatusova T, Thomson N, Ashburner M, Boore JL, Cochrane G, Cole J: Towards richer descriptions of our collection of genomes and metagenomes. 2008
- Bossy R, Jourde J, Manine A-P, Veber P, Alphonse E, Van De Guchte M, Bessières P, Nédellec C: BioNLP 2011 Shared Task - The Bacteria Track. BMC Bioinformatics. 2012, 13: S8-View Article
- Bossy R, Jourde J, Bessieres P, van de Guchte M, Nédellec C: BioNLP Shared Task 2011-Bacteria Biotope. ACL HLT 2011; Portland, Oregan. 2011, 56-
- Kim JD, Ohta T, Pyysalo S, Kano Y, Tsujii J: Extracting bio-molecular events from literature - the BioNLP'09 shared task. Computational Intelligence. 2011, 27: 513-540. 10.1111/j.1467-8640.2011.00398.x.View Article
- GuoDong Z, Jian S, Jie Z, Min Z: Exploring various knowledge in relation extraction. Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics: June 2005; Ann Arbor. 2005, 427-434.
- Kripke SA: Naming and necessity. 1981, Wiley-Blackwell
- Nédellec C, Golik W, Aubin S, Bossy R: Building large lexicalized ontologies from text: a use case in automatic indexing of biotechnology patents. Knowledge Engineering and Management by the Masses. 2010, 514-523.View Article
- NCBI taxonomy. [http://www.ncbi.nlm.nih.gov/Taxonomy/]
- Lapage SP, Sneath PHA: International code of nomenclature of bacteria. 1992, ASM Press
- GEOnet Names Server (GNS). [http://earth-info.nga.mil/gns/html/]
- Agrovoc. [http://aims.fao.org/standards/agrovoc/about]
- DBpedia. [http://dbpedia.org/About]
- Aubin S, Hamon T: Improving term extraction with terminological resources. Advances in Natural Language Processing. 2006, 380-387.View Article
- Liolios K, Chen IMA, Mavromatis K, Tavernarakis N, Hugenholtz P, Markowitz VM, Kyrpides NC: The Genomes On Line Database (GOLD) in 2009: status of genomic and metagenomic projects and their associated metadata. Nucleic acids research. 2010, 38: D346-D354. 10.1093/nar/gkp848.PubMed CentralView ArticlePubMed
- Jacquemin C, Tzoukermann E: NLP for term variant extraction: synergy between morphology, lexicon and syntax. Natural language information retrieval. 1999, 7: 25-74.View Article
- Hamon T, Nazarenko A: Detection of synonymy links between terms. Recent Advances in Computational Terminology. 2001, 185-208.View Article
- Aronson AR: Effective mapping of biomedical text to the UMLS Metathesaurus: the MetaMap program. Proceedings of the AMIA Symposium: 3-7 Nov 2001. 2001, Washington, DC, USA, 17-
- Fundel K, Küffner R, Zimmer R: RelEx--Relation extraction using dependency parse trees. Bioinformatics. 2007, 23: 365-10.1093/bioinformatics/btl616.View ArticlePubMed
- Manine AP, Alphonse E, Bessières P: Learning ontological rules to extract multiple relations of genic interactions from text. International journal of medical informatics. 2009, 78: e31-e38. 10.1016/j.ijmedinf.2009.03.005.View ArticlePubMed
- Zelenko D, Aone C, Richardella A: Kernel methods for relation extraction. The Journal of Machine Learning Research. 2003, 3: 1083-1106.
- Culotta A, McCallum A, Betz J: Integrating probabilistic extraction models and data mining to discover relations and patterns in text. Proceedings of the main conference on Human Language Technology Conference of the North American Chapter of the Association of Computational Linguistics(HLT-NAACL); New York City. 2006, 296-303.View Article
- Airola A, Pyysalo S, Björne J, Pahikkala T, Ginter F, Salakoski T: A graph kernel for protein-protein interaction extraction. Proceedings of the Workshop on Current Trends in Biomedical Natural Language Processing:19 June 2008; Columbus, Ohio. 2008, 1-9.View Article
- Bollegala D, Matsuo Y, Ishizuka M: Relation Adaptation: Learning to Extract Novel Relations with Minimum Supervision. Twenty-Second International Joint Conference on Artificial Intelligence: 16-22 July 2011; Barcelona. 2011
- Kambhatla N: Combining lexical, syntactic, and semantic features with maximum entropy models for extracting relations. Proceedings of the ACL 2004 on Interactive poster and demonstration sessions; Barcelona. 2004, 22-View Article
- Ono T, Hishigaki H, Tanigami A, Takagi T: Automated extraction of information on protein-protein interactions from the biological literature. Bioinformatics. 2001, 17: 155-10.1093/bioinformatics/17.2.155.View ArticlePubMed
- Torii M, Vijay-Shanker K: Sortal Anaphora Resolution in Medline Abstracts. Computational Intelligence. 2007, 23: 15-27. 10.1111/j.1467-8640.2007.00292.x.View Article
- Lappin S, Leass HJ: An algorithm for pronominal anaphora resolution. Computational Linguistics. 1994, 20: 535-561.
- Segura-Bedmar I, Crespo M, De Pablo-Sánchez C, Martínez P: Resolving anaphoras for the extraction of drug-drug interactions in pharmacological documents. BMC Bioinformatics. 2010, 11 (Supl 2): S1-PubMed CentralView ArticlePubMed
- Lin Y, Liang T: Pronominal and sortal anaphora resolution for biomedical literature. Proceedings of ROCLING XVI: Conference on Computational Linguistics and Speech Processing; Taipei. 2004
- Grosz BJ, Weinstein S, Joshi AK: Centering: A framework for modeling the local coherence of discourse. Computational linguistics. 1995, 21: 203-225.
- Erkan G, Ozgur A, Radev DR: Semi-supervised classification for extracting protein interaction sentences using dependency parsing. Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL): 28-30 June 2007; Prague. 2007, 1: 228-237.
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.