Skip to main content
  • Methodology Article
  • Open access
  • Published:

Biomedical event extraction with a novel combination strategy based on hybrid deep neural networks

Abstract

Background

Biomedical event extraction is a fundamental and in-demand technology that has attracted substantial interest from many researchers. Previous works have heavily relied on manual designed features and external NLP packages in which the feature engineering is large and complex. Additionally, most of the existing works use the pipeline process that breaks down a task into simple sub-tasks but ignores the interaction between them. To overcome these limitations, we propose a novel event combination strategy based on hybrid deep neural networks to settle the task in a joint end-to-end manner.

Results

We adapted our method to several annotated corpora of biomedical event extraction tasks. Our method achieved state-of-the-art performance with noticeable overall F1 score improvement compared to that of existing methods for all of these corpora.

Conclusions

The experimental results demonstrated that our method is effective for biomedical event extraction. The combination strategy can reconstruct complex events from the output of deep neural networks, while the deep neural networks effectively capture the feature representation from the raw text. The biomedical event extraction implementation is available online at http://www.predictor.xin/event_extraction.

Background

PubMed recorded over 28 million papers in 2018 [1] which reflects the rapid growth of the biomedical literature. The knowledge and discoveries reported in the biomedical literature receive substantial attention, but the large volume of the literature poses a challenge to information retrieval; therefore, text mining has become an in-demand technology and a popular research focus. Event extraction, which is an effective way to represent the structured knowledge from unstructured text [2], is a fundamental technology for text mining. However, event extraction is particularly difficult due to the complex and arbitrary structure of events in biomedicine, so related research is urgently needed [3].

The definition of a biomedical event, according to the BioNLP [4], consists of (1) a trigger word that indicates the existence of an event and belongs to a certain event type and (2) multiple arguments in which an argument can be viewed as a relation between the event triggers and entities or other event, and each argument has an argument type as well. Therefore, the task of event extraction is to recognize the event triggers with their arguments from the raw text.

We illustrate biomedical events with Fig. 1 as example. The word “promote” is an event trigger of the event type Positive Regulation. This event has a Theme argument linked to the word “tumorigenesis”, which is an entity of Carcinogenesis type, and an Cause argument linked to “over-expression”. Notice that some events can be the argument for other events, i.e., a nested structure, such as “over-expression” serving as an Gene Expression event trigger as well as an argument of Positive Regulation event. Therefore, the event can be viewed as a directed graph in the text for which the node of graph is the event trigger or entity and the directed edges indicate the arguments.

Fig. 1
figure 1

A sentence with visualized events, provided by BioNLP-ST2013

Since biomedical event extraction was defined as a standard task, various methods have been proposed. Most previous work can be classified into three types: rule-based approaches [5, 6], traditional shallow machine learning models and deep learning models. The Turku Event Extraction System (TEES) [7, 8] is a biomedical event extraction system that uses rich features from dependency parsing. The TEES utilizes a step-wise approach based on multi-class SVMs by breaking down the whole task into straightforward consecutive graph node and edge classification tasks. The EventMine [9] is a similar SVM-based pipeline method with handcrafted features. Majumder et al. [10] exploited a stacking model for biomedical event extraction. The system uses SVC, SGD and LR as the base-level classifiers and takes SVC as the meta-level classifier. A transition-based model for event extraction [11] is another approach leveraging a structured perception for encoding and decoding with a beam search to find the global best prediction.

In recent years, deep learning methods have been applied to this task that extend the feature representation from text and promote the performance. Wang et al. [12] proposed a convolutional neural network (CNN) with multiple distributed features for biomedical event extraction. The distributed features contain not only the word embedding but also trigger types, POS labels and topic representation. Li et al. [13] utilized dependency-based word embedding and a parallel multi-pooling convolutional neural network to extract biomedical events. This approach reserves more information by pooling the multi-segment of a sentence divided by word triggers and arguments. Björne and Salakoski [14] integrated CNN into the original TEES to supply more features, and replaced the SVM classifier with dense layers, which suggested that the inclusion of the neural network significantly enhanced the performance. Li et al. [15] proposed a framework that using gated recurrent unit networks with attention mechanism to extract biotope and bacteria events.

However, deep learning methods are still rarely used for biomedical event extraction, which is partially due to the complexity of the task-specific event structures. More deep models have focused on the sub-tasks of event extraction such as event trigger detection [1618] and relation classification [1922], and most of these models obtained superior performance compared to traditional shallow methods.

Despite the success of existing methods in biomedical event extraction, they generally suffer from two limitations. First, most of them heavily rely on manually designed features and usually need complicated natural language processing (NLP) from external NLP toolkits with poor generalizability. Second, these methods organize the task in a pipeline manner and separate it into independent sub-tasks, which simplifies the problem but ignores the interaction between the sub-tasks and makes the process prone to accumulating errors.

Due to the aforementioned limitations, we propose a biomedical event extraction method with a novel combination strategy based on deep neural networks. Our method detects the candidate event triggers and relations from raw text with recurrent neural networks (RNN) and convolutional networks (CNN), and then the Combination Strategy (CS) constructs the event from the detected results by solving an optimization problem. The proposed method takes advantage of neural networks that can represent features from word embedding in semantic space [23] and removes the reliance on feature engineering. The CS, which integrates global information to optimize a penalty, alleviates the error accumulation.

We evaluated our method with three common biomedical event extraction tasks: the Multi-Level Event Extraction (MLEE) [24], Cancer Genetics (CG) and Pathway Curation (PC) from BioNLP Shared Task 2013 (BioNLP-ST2013) [4]. Our method outperforms the state-of-the-art methods for all of these tasks according to overall F1 scores. The experimental results demonstrate the effectiveness and generalizability of the hybrid networks and the CS. Additionalty, our method only needs a minimized task-specific configurations without the adjustment on method, which makes it easy to facilitate for various biomedical event extraction tasks.

The contributions of this paper are summarized below:

  • Describes the first attempt to use hybrid deep neural networks (CNN and RNN) aimed at achieving biomedical event extraction.

  • Proposes a novel combination strategy to integrate the detected triggers and relations in an optimized manner.

  • Utilizes end-to-end learning as well as avoids reliance on manual feature design and external NLP packages.

Results

Dataset

We trained and evaluated our method using three common annotated datasets: CG, PC and MLEE. Each dataset is initially divided into three parts: training, development and test sets, and the statistics of these datasets are listed in Table 1. CG data concerns the extraction of events relevant to cancer, including molecular foundations, cellular tissue, and organ-level effects. PC data targets reactions relevant to the development of biomolecular pathway models. MLEE data focuses on events across multiple levels of biological organization from the molecular level to the organ system level. All of these datasets have provided entity labels for each word so that task can focus on targeting event extraction.

Table 1 Statistics of datasets

The pre-processing is simple; we only split each document into sentences and tokenized them into sequences of words, which does not rely on any additional NLP toolkits.

Training

The main hyper-parameters, which were tuned on the development set, were set as follows: learning rate = 0.007, ratio of class weight of positive and negative classes for TR and RC was 5:1, weight decay = 0.0002, batch size = 16, α = 0.5, β = 0.25, γ = 0.125, thresholdt=−2.0 and thresholdr=−2.0. The other hyper-parameters of our model are listed in Additional file 1: Table S11. We set k=10 in inverse sigmoid decay function. The activation function was leaky-relu. The optimizer we used was Adam. We used 5 single models for ensemble learning. Some tricks were employed, including using pretrained Char-level CNN and word embedding on large external corpus [25], Xavier initialization for neural layers [26], dropout in LSTM and undersampling in EE. The undersampling is to keep the class balance of positive/negative event samples in each sentence.

The final model were trained by the union of training and development set through 100 epochs for each task and the loss curve on CG corpus is shown as Fig. 2. The loss of each module was calculated individually and the gradient of them was propagated simultaneously to update parameters at the end of every batch. The loss of TR declines first, followded by the loss of RC, while the loss of EE has the slowest change. This phenomenon is reasonable because the latter two losses rely on the former detected results. The black curve is the total loss that sums up of loss from TR, RC and EE. The oscillation of the loss curve is due to the variance in the length of sentences and different number of events among the batches. The figure shows that the model has converged after 100 training epochs. The 100-epochs training on CG corpus (400 documents) takes about 12 h on an i7-7700 CPU. The prediction of a single model for each document takes about 20 s on average and ratio of time consuming of each module (TR, RC, EE and CS) is 55.97%, 20.75%, 23.06% and 0.23%, respectively.

Fig. 2
figure 2

The sequence of training loss within 100 epochs

Performance

We used standard recall, precision and F1 scores as evaluation metrics. The event was regarded as true-positive only if both trigger and arguments were detected correctly. Our evaluation followed the primary criteria, i.e., approximate span matching and approximate recursive matching [27].

Table 2 shows the overall performance of our method and other state-of-the-art methods for CG, PC and MLEE on test set. RelAgent [6] is a linguistically motivated rule-based system to extract biomedical events. NCBI [5] uses an approximate sub-graph matching-based approach. Zhou and Zhong [28] utilized a semi-supervised learning framework with un-annotated corpora. The TEES [7, 8] and EventMine [9] are both SVM-based pipeline models with hand-designed rich features. The TEES CNN [14] is the upgraded version of TEES coupled with CNN and uses mixed 5 model ensemble with randomized train/development set split. Wang et al. [12] and Li et al. [13] both developed convolutional network-based methods.

Table 2 Comparison of overall performance on CG, PC and MLEE task (test set)

Table 2 shows that our method achieved the highest F1 scores for all three datasets by 58.04% for CG, 55.73% for PC and 60.05% for MLEE respectively, which suggests the effectiveness and generalization ability of the hybrid networks and CS. We conducted student’s t-test [29] on the best existing F1 score and F1 scores of our proposed method in multiple runs. The results indicate that the improvements are statistically significant on CG and MLEE task with p-value <10−3, and p-value on PC task is 0.062 (the detailed statistics are listed in Additional file 1: Table S7–S9). The precision was dramatically higher than the recall in all datasets, which was probably due to the highly diverse event schemes and insufficient training set.

Table 3 shows the detailed performance for our method and other existing methods for grouped event categories of CG. As shown in Table 3, our method outperformed other methods for all the event categories. The most significant improvement was for Modification events, and these events relied highly on the global contextual information that was modelled precisely by the recurrent network of EE. Through vertical comparison, the lower scores for Regulation, Planned Pro and Modification compared to those of other categories were due to their nested structure, i.e., these events usually took other events as arguments, which were more difficult to correctly detect. The full detailed performance for CG, PC and MLEE is listed in Additional file 1: Table S4–S6.

Table 3 Detailed performance comparison on CG

Discussion

Alternatives comparison

Before the proposed method was determined, we conducted ablation experiments on several variations of the proposed method to validate the effectiveness of each part of the method. Table 4 shows results of the ablation study, which evaluated the ensemble learning, the EE module, CS algorithm, threshold setting and Char-CNN module. In these experiments, methods were trained on training set and tested on development set. The Single-model is the proposed method in singleton without ensemble learning. The Single-pipeline-model seperates TR, RC and EE into independent networks without parameter sharing. The Combination-rule-single does not use the EE module and replaces the CS with a rule-based method to assign only one event to each detected event trigger. The Combination-rule-all similarly replaces the CS with another rule-based method to generate all possible combinations from triggers and relations. The EE-probability directly uses the probability outputing from TR, RC and EE to determine the final events instead of using CS, i.e., the method extracts events simply by the classification results of each network modules (setting all thresholds to 0). The Zero-threshold resets both thresholdt and thresholdr to zero and keeps other setting same as proposed method. The Without-CharCNN removes the Char-CNN module from the method. The Single-model-pipeline and Without-CharCNN need retrain the neural networks while other methods do not retrain the networks since they only change the settings in post-processing steps.

Table 4 Performance comparison across variations of our method on development set

As shown in Table 4, the proposed method achieved higher F1 scores than other alternative variations. The contrast results of Single-model show that the ensemble learning noticeably improved the F1 performance by 3%-5%, which was mainly contributed by the improvement in precision of 6%-8%. Since the randomness in the training led to variance among the different runs (which also caused performance differences by several percentage points, e.g., the F1 scores by multiple runs of Single-model on CG have a standard deviation of 0.532%), the ensemble could eliminate the variance and obtain higher precision. The F1 scores of Single-model-pipeline sightly declined by 0.2%-1.6% that suggests parameter sharing is effective and efficient (the training of Single-model-pipeline takes longer time than proposed joint model). The performance of Combination-rule-single decreased markedly by 3%-6% compared to the proposed method, especially for recall because the method only assigned one event for each detected trigger and multiple events associated with one trigger were discarded through its rule (such multiple events accounts for 59.7% of total events in CG corpus, for example). The Combination-rule-all obtained high recall that is even higher than the recall of our proposed method on CG and PC tasks, but it suffered from much lower precision and the F1 scores decreased markedly because it constructed too many incorrect events. Both of these ablation tests show that the EE module is valuable. Removing EE module yeilded a 0.8%-2.6% decline of F1 score by EE-probability, which indicates that the CS contributes a positive effect by optimizing the penalty value. However, EE-probability obtained higher precision because the method applied stricter extraction. EE-probability assigned an event only when all of TR, RC and EE modules returned positive classification results, thus the detected events were lesser then the proposed method did (e.g., 2225 vs. 2354 on CG development set). Additionally, the comparison with the Zero-threshold demonstrates that setting a lower threshold for the candidate triggers and relations can construct more valid events and promote overall performance. The contrast results in the last row demonstrate that the Char-CNN module is beneficial to the overall performance by providing the lexical features.

Error and limitation analysis

We divided extraction errors into five types, including wrong trigger label, wrong trigger span, wrong arguments, redundant arguments and other errors. Detailed statistics for the errors in prediction phase are listed in Table 5. The statistics suggested that the most common error type was the wrong trigger span, which constituted about half of the total errors, indicating that the range of trigger words is the most difficult information to detect.

Table 5 Statistics for the extraction errors in CG/PC/MLEE

Moreover, similar to other works, two special cases were ignored to simplify in our method. First, a few events had trigger words and arguments spanning across more than one sentence, but our method only detects events within a single sentence. Secondly, a few words were associated with more than one event trigger labels, but our method could assign only one trigger label to them. Ignoring such cases could cause a performance reduction, but these cases are relatively rare (approximately 2% - 4.5%) and had limited effect (see Additional file 1: Table S10).

Some limitations still exist and need further improvement. Our method is based on a deep neural network with a large number of learnable parameters but the training set with hundreds of documents is somewhat insufficient. The tuning of hyper parameters relied on grid searching in a development set, which was time consuming.

Conclusions

In this paper, we present deep neural networks coupled with a combination strategy to extract biomedical events. Our method detects the event trigger and classifies the relations jointly while taking advantage of deep neural networks that extract feature representation automatically and do not rely on manual feature engineering. This novel Combination Strategy integrates the outputs from different stages to construct the events in an optimization manner, which alleviates the error accumulation. The evaluation results show that our method has achieved state-of-the-art performance compared to existing methods, which indicates that the Combination Strategy and the deep neural networks in our method are effective. In the future, we plan to extend our method with semi-supervised learning to address the insufficiency of the training corpora. Since biomedical text mining is a desirable technology for converting the large number of articles to structured information at high-layer semantics, we believe the proposed method has the potential to facilitate event extraction in broad, real-world scenarios for researchers.

Methods

We apply end-to-end supervised deep learning for event extraction. The overall architecture of the networks is illustrated in Fig. 3, which consists of 5 modules: the Character-level CNN (CharCNN), Bi-directional LSTM (BiLSTM), Trigger Recognition (TR), Relation Classification (RC) and Event Evaluation (EE). The CharCNN and BiLSTM encode a sentence into a sequence of feature vectors. The TR, RC and EE are stacked on BiLSTM and determine the type and probability of each event trigger, relation and candidate event, respectively. These modules are trained simultaneously in a joint manner, which can benefit from parameter sharing [19, 21]. Finally, the outputs of these modules are integrated into the CS, which is a post-processing step that is applied in prediction phase to generate the final events.

Fig. 3
figure 3

The overall networks structure (CharCNN, BiLSTM, TR, RC, EE), some components are omitted for brevity(detailed structures are shown in Additional file 1: Figure S1–S3)

We cast the argument assignment as a relation classification task, so we use the term relations instead of arguments in the rest of this paper.

Character-level cNN

Character-level CNN (CharCNN) extracts the character-level features of each word. The module is inspired by previous work [21] that has been shown to be effective due to the ability to capture the morphological information [30].

For each word, the module first looks up an embedding layer to get the vector representation of each character. Let the sequence of embedding vectors be \(V^{(c)} = \left \{v^{(c)}_{1,}v^{(c)}_{2},...,v^{(c)}_{n}\right \}\) where \(v^{(c)}_{i}\) is the vector of i-th character. Then, the sequence is fed to convolution layer, which is computed by:

$$ y^{(c)}_{i} =f\left(conv\left(W_{1},V^{(c)}_{i:i+k}\right)+b_{1}\right), $$
(1)

where k denotes the kernel size, conv(·,·) is the convolutional operator, f(·) is the activation function, \(W_{1}\in \mathbb {R}^{d\times k\times nc}\) is the parameter of the convolution layer where nc is the number of output channels and \(b_{1}\in \mathbb {R}^{nc}\) is the bias vector. Therefore, we obtain the representation matrix \(y^{(c)}\in \mathbb {R}^{n\times nc}\) for an n-length word. To obtain the fixed length representation of the word, the adaptive max pooling is then applied to the output vector:

$$ ch_{j} = \max_{1\leq i\leq n}y^{(c)}_{ij}, $$
(2)

where \(ch\in \mathbb {R}^{nc}\) is the char-level representation of the word.

Bi-directional lSTM

The Bi-directional LSTM (BiLSTM) encodes a sentence into a list of hidden vectors. The LSTM can model the long-distance dependency that benefits from its memory and forget blocks, and a signal from two directions helps the module sense the context [31].

Given a sentence with n words, the word embedding layer maps each word into a vector as wi. Similarly, the entity label of each word is also mapped to vector ei by the entity label embedding layer. We have obtained the character-level representation of each word denoted as chi from Char-CNN. Then, the above vectors are concatenated and denoted as vi=[wi,ei,chi]. The vector representation of n words forms V={v1,v2,...,vn}, and then it is fed to the LSTM layers with two parallel (forward and backward) directions. The computation of the LSTM layer at the time step i is:

$$ \begin{aligned} & i_{i} = \sigma (W_{jj}v_{i}+b_{jj}+W_{hj}h_{i-1}+b_{hj}), \\ & f_{i} = \sigma (W_{jf}v_{i}+b_{jf}+W_{hf}h_{i-1}+b_{hf}), \\ & g_{i} = \tanh(W_{jg}v_{i}+b_{jg}+W_{hg}h_{i-1}+b_{hg}), \\ & o_{i} = \sigma (W_{jo}v_{i}+b_{io}+W_{ho}h_{i-1}+b_{ho}), \\ & c_{i} = f_{i}\cdot c_{i-1}+i_{i}\cdot g_{i}, \\ & h_{i} = o_{i}\cdot \tanh(c_{i}), \end{aligned} $$
(3)

where ii,fi,gi,oi and ci are the input gate, forget gate, intermediate state, output gate and cell state, respectively. σ(·) is the sigmoid function. \(h_{i} = \left [\overrightarrow {h_{i}},\overleftarrow {h_{i}}\right ]\) is the hidden vector from LSTM at time step i, which consists of two directions. Finally, we obtain the sentence encoding sequence \(H = \{h_{1},h_{2},...,h_{n}\}\in \mathbb {R}^{2hd\times n}\) where hd is the hidden size. Additionally, we also obtain the sequence of entity label embedding E={e1,e2,...,en}.

Trigger recognition

We cast the Trigger Recognition (TR) as a sequence labelling task. The TR module receives the output of BiLSTM and assigns an event label to each word in the sequence in the BILOU scheme [32].

Given the input sequence H={h1,h2,...,hn}, we assign the label in a greedy manner from left to right. At the time step i, we concatenate the encoded vector hi and previous event trigger label vector ti−1 into xi=[hi,ti−1] and then send them into a linear layer and a log softmax layer, which is written as:

$$ \begin{aligned} & y^{(t)}_{i} =W_{3}(f(W_{2}x_{i}+b_{2})+b_{3}), \\ & p^{(t)}_{i,j} =\log\left(\frac{\exp\left(y^{(t)}_{i,j}\right)}{\sum_{k} \exp\left(y^{(t)}_{i,k}\right)}\right), \end{aligned} $$
(4)

where W2 and W3 are the weight matrices of the two linear layers, respectively, b2 and b3 are the bias vectors, and f(·) is the activation function. The softmax layer transforms the \(y^{(t)}_{i}\) to the trigger label probability vector \(p^{(t)}_{i}\). The event trigger label of the i-th word is assigned as the m-th trigger type where m is the index of the maximal element in \(p^{(t)}_{i}\) except for \(p^{(t)}_{i,none}\) if \(p^{(t)}_{i,m}-p^{(t)}_{i,none}>threshold_{t}\); otherwise we assign none to the word. The trigger label is then transformed to ti by the event trigger label embedding layer and then sent to the next time step. Finally, we obtain the sequence of trigger label embedding vectors T={t1,t2,...,tn}.

Here we define a support value to measure the confidence of the assigned label.

Definition 1

A Support Value is the probability difference between the assigned label m and the label none. The lager support value means the more confidence of the recognised label.

For each recognized event trigger with index m, a support value is computed by:

$$ s^{(t)} = p^{(t)}_{m}-p^{(t)}_{none}, $$
(5)

where none is the index of none type.

In the training of TR, we use a scheduled sampling trick to eliminate the gap between training anf inference [33]. We take inverse sigmoid decay function ε=k/(k+ exp(i/k)) to decide the probability of using true token or inference token, where i is the number of training epochs.

Relation classification

The Relation Classification (RC) module predicts the relation type for each candidate pair. The sentence representation is derived from BiLSTM, and the event trigger is detected by the TR before. Here, we use the information from the events/entities and the sub-sentence between them to predict the type of relation.

Since the TR module has detected the event triggers and the entities are given, we combine all possible trigger-entity and trigger-trigger pairs. The set of candidate pairs is written as RL={(e(1),e(2))|e(1)event_trigger_set,e(2)event_trigger_setentity_set}. To predict the relation type of (e(1),e(2))RL, we first take the truncated sequence of BiLSTM hidden vectors \(H_{start_{1}:end_{1}}\), the sequence of trigger label vectors \(T_{start_{1}:end_{1}}\) and entity label vectors \(E_{start_{1}:end_{1}}\), where start1,end1 are the start and end location of e(1). The three vectors are concatenated into \(src = [H_{start_{1}:end_{1}}, T_{start_{1}:end_{1}}, E_{start_{1}:end_{1}}]\). Then, we get \(dst = [H_{start_{2}:end_{2}}, T_{start_{2}:end_{2}}, E_{start_{2}:end_{2}}]\) in same manner for e(2). The sub-sentence between e(1) and e(2) is denoted as \(mid = [H_{end_{1}:start_{2}}, T_{end_{1}:start_{2}}, E_{end_{1}:start_{2}}]\). Here, we assume end1<start2 without loss of generality. If e(1) and e(2) are adjacent, we assign a learnable vector to mid to represent this situation.

The src and dst are then fed to the adaptive max pooling layer to get a vector with a fixed shape that is denoted as srcmax and dstmax, which is the same as Eq. 2. Meanwhile, the sub-sentence mid is fed to a convolutional layer and an adaptive max pooling layer to extract the feature vector, which is denoted as midmax, same as Eq. 1.

Additionally, Zheng et al. [20] mentioned that the distance of two entities start2end1 can determine the relation significantly. Therefore, we use a distance embedding vector to provide extra information that is ignored by the max pooling operation, and the vector is denoted as dv.

The above vectors are concatenated into r=[srcmax,midmax,dstmax,dv] and apply two linear layers:

$$ y^{(r)} = W_{5}(f(W_{4}r+b_{4})+b_{5}), $$
(6)

where W4,W5 and b4,b5 are the weight matrices and bias vectors of two linear layers, respectively. Finally, the vector is fed into a log softmax layer as follows:

$$ p^{(r)}_{j} =\log\left(\frac{\exp\left(y^{(r)}_{j}\right)}{\sum_{k} \exp\left(y^{(r)}_{k}\right)}\right), $$
(7)

where \(p^{(r)}\in \mathbb {R}^{c}\) is the probability that the relation belongs to each relation class of total c classes. We assign the m-th class of relation to the candidate pair (e(1),e(2)) where m is the index of the maximal element of p(r) except for \(p^{(r)}_{none}\) if \(p^{(r)}_{m}-p^{(r)}_{none}>threshold_{r}\); otherwise we assign none to the relation.

For each recognized relation with class index m, we compute a support value by:

$$ s^{(r)} = p^{(r)}_{m}-p^{(r)}_{none}, $$
(8)

where none is the index of none type.

Event evaluation

Since the entities, event triggers and relations have been recognized, we then apply the Event Evaluation (EE) module to estimate the probability that a candidate event structure is a valid event.

We use another BiLSTM to represent the event structure. We have obtained the word encoding sequence H={h1,h2,...,hn} and the entity label embedding sequence E={e1,e2,...,en} from BiLSTM, the trigger label embedding sequence T={t1,t2,...,tn} from TR, and the recognized relation set \(RL\_recognized = \left.\left \{\left (e^{(1)}_{i},relation\_type_{i},e^{(2)}_{i}\right)\right |i=1,2,... \right \}\) from RC. Then, we enumerate all the valid combinations of event triggers and relations that have to meet the structure definition of the certain task. All the valid combinations in a sentence form the candidate events set denoted as C.

For each candidate event in C, a sequence of role label R={r1r2,...,rn} is leveraged to represent the role of each word in the candidate event. The i-th item in R is assigned to event_trigger or relation_typek or none_type according to the role that the i-th word plays. The role label sequence is then transformed into vectors by the role label embedding layer, and the result is denoted as \(V^{(e)} = \left \{v^{(e)}_{1},v^{(e)}_{2},...,v^{(e)}_{n}\right \}\).

Then, we concatenate each corresponding vector from sequence H, E, T and V(e) in parallel, which is written as X={x1,x2,...,xn} where \(x_{i} = [h_{i},e_{i},t_{i},v^{(e)}_{i}]\). The sequence X is fed to another BiLSTM layer:

$$ h^{(e)}_{i}=BiLSTM(x_{i}), i=1,2,...,n. $$
(9)

The last outputs from two directions of BiLSTM \(h^{(e)}_{last}=\left [\overrightarrow {h^{(e)}_{n}},\overleftarrow {h^{(e)}_{1}}\right ]\) are fed to a full connection layer and then mapped to log probability \(p^{(e)}\in \mathbb {R}^{2}\) by a log softmax layer. The probability vector p(e) denotes that the event is a positive sample or a negative sample. Meanwhile, we use another full connection layer and log softmax layer, which is branched from \(h^{(e)}_{last}\), to learn a vector \(p^{(m)} \in \mathbb {R}^{3}\) that denotes the modification information of the event. The three items in p(m) represents the log-probability of Negation, Speculation and None, respectively.

For each candidate event, we compute a support value by:

$$ s^{(e)} = p^{(e)}_{1}-p^{(e)}_{2}, $$
(10)

where \(p^{(e)}_{1}\) is the log probability that the candidate event is a valid event and \(p^{(e)}_{2}\) is not.

After we obtain the probability vectors p(t),p(r),p(e) and p(m) from TR, RC and EE, an end-to-end training is then applied to optimize the NLLloss of these vectors according to the correct labels of triggers, relations, events and modifications, respectively.

Combination strategy

After we have obtained the support value of each candidate trigger, relation and event, the Combination Strategy (CS) is then applied to integrate all of the information and determine the set of final extracted events in prediction phase. CS does not participate in training.

The principle of CS is to minimize the penalty value, which is designed to measure the discordance between the final events and the outputs of previous modules. The penalty consists of two parts as follows. If a trigger or a relation has a positive support value, but it does not appear in any event of the final extracted events, it generates a penalty called “support waste”. In contrast, if a candidate event with a negative support value is included as one of the final extracted events, it generates a penalty called “support lacking”. These two penalties are conflicting objectives. The more candidate events that are added into the final set, the less “support waste” penalty will be, but the penalty for “support lacking” will increase, and vice versa. So, the goal will be to minimize the total penalty so that the final extracted events can be determined.

Formally, this target is written as follows:

$$ c_{best}=\mathop{\arg\min}_{c\subseteq C} penalty\_score(c), $$
(11)

where cbest is the set of final extracted events, c enumerates all the subsets of C, and:

$$ \begin{aligned} penalty\_score(c)=\left(\sum_{event_{k}\in c}\max(1-\alpha\cdot s^{(e)}_{k},0)+ \right.\\ \left.\beta\sum_{trigger_{i}\notin c}s^{(t)}_{i}+\gamma\sum_{relation_{j}\notin c}s^{(r)}_{j}\right), \end{aligned} $$
(12)

where \(s^{(e)}_{k}\) denotes the support of the k-th candidate event, \(s^{(t)}_{i}\) denotes the support of i-th event trigger, and \(s^{(r)}_{j}\) denotes the support of the j-th relation. The first term accounts for the support lacking penalty, the last two terms accounts for the support waste penalty and reweighted by parameters α, β and γ.

Figure 4 is a simple example to illustrate the principle of CS. As shown in Fig. 4, three entities (T1, T3 and T5) were given, and two event triggers (T2 and T4) and six relations were detected, then a total of six candidate events were constructed from them. After computing them, the minimized penalty was obtained when choosing the two of them to be the final result.

Fig. 4
figure 4

A simple example to illustrate the principle of CS

To optimize Eq. 11, we have to enumerate all of the subsets of C, which requires exponential time complexity. For some sentences, there are many complex events with many candidate arguments and lead to a very large computation cost (26 in Fig. 4 for example). Therefore, the CS, which is an approximation algorithm, is proposed to solve the problem within the O(n2) time complexity.

The CS receives the set of candidate events C, the support value of triggers, relations and candidate events in a sentence, and then it returns the set of final extracted events. Before the CS, the candidate events are sorted by the topological order of their nested structure if it exist, and the events relating to different triggers are handled independently, except for transmitting the support value between them. In the CS, the initial set of chosen events is empty, and then the candidate events are added into the set one by one in a greedy manner. The pseudo code of the CS is shown in Algorithm 1.

To prevent the nested events from forming event loops, we set a loop detector after CS. The extracted events are added into the final set one-by-one, we discard the event if the inclusion of it would cause event loop. We assign event modifications (sepculation/negation or none-modification) for the events in the final extracted set according to their modification vectors p(m).

There are two advantages of the CS. First, the CS can address event triggers and relations with negative support value. In the TR and RC, we can set negative thresholds thresholdt and thresholdr to obtain more candidate event triggers and relations, which is especially useful for ensemble learning to improve recall. Second, the CS can alleviate the error accumulation effectivelly. Even though a true event trigger is wrongly assigned a negative support value (TR fails), but the related arguments and event structure are recognized as positive instances (RC and EE work), the trigger can still be correctly constructed as a final event with the CS.

Availability of data and materials

The source code is available at https://github.com/melissa135/Event_Extraction. The biomedical event extraction implementation is available online at http://www.predictor.xin/event_extraction. The datasets that used in experiments are available online at http://2013.bionlp-st.org/tasksand http://nactem.ac.uk/MLEE/.

Abbreviations

CG:

Cancer genetics

CNN:

Convolutional neural network

CS:

Combination strategy

EE:

Event evaluation

LR:

Linear Regression

LSTM:

Long-short term memory

MLEE:

Multi-level event extraction

NLP:

Natural language processing

PC:

Pathway curation

POS:

Part of speech

RC:

Relation classification

RNN:

Recurrent neural network

SGD:

Stochastic gradient descent

SVC:

Support vector classification

SVM:

Support vector machine

TR:

Trigger recognition

References

  1. Fiorini N, Canese K, Starchenko G, Kireev E, Kim W, Miller V, et al.Best Match: new relevance search for PubMed. PLoS Biol. 2018; 16(8):e2005343.

    Article  Google Scholar 

  2. Huang CC, Lu Z. Community challenges in biomedical text mining over 10 years: success, failure and the future. Brief Bioinforma. 2015; 17(1):132–44.

    Article  Google Scholar 

  3. Ananiadou S, Pyysalo S, Tsujii J, Kell DB. Event extraction for systems biology by text mining the literature.Trends Biotechnol. 2010; 28:381–90.

    Article  CAS  Google Scholar 

  4. Pyysalo S, Ohta T, Rak R, Rowley A, Chun HW, Jung SJ, et al.Overview of the cancer genetics and pathway curation tasks of bionlp shared task 2013. BMC Bioinforma. 2015; 16(10):S2.

    Article  Google Scholar 

  5. Liu H, Verspoor K, Comeau DC, MacKinlay AD, Wilbur WJ. Optimizing graph-based patterns to extract biomedical events from the literature. BMC Bioinforma. 2015; 16(16):S2.

    Article  Google Scholar 

  6. Ramanan S, Nathan PS. Performance and limitations of the linguistically motivated Cocoa/Peaberry system in a broad biological domain: Citeseer; 2013, p. 86.

  7. Björne J, Heimonen J, Ginter F, Airola A, Pahikkala T, Salakoski T. Extracting contextualized complex biological events with rich graph-based feature sets. Comput Intell. 2011; 27(4):541–57.

    Article  Google Scholar 

  8. Björne J, Salakoski T. TEES 2.2: biomedical event extraction for diverse corpora. BMC Bioinforma. 2015; 16(16):S4.

    Article  Google Scholar 

  9. Miwa M, Ananiadou S. Adaptable, high recall, event extraction system with minimal configuration. BMC Bioinforma. 2015; 16(10):S7.

    Article  Google Scholar 

  10. Majumder A, Ekbal A, Naskar SK. Biomolecular Event Extraction using a Stacked Generalization based Classifier. In: Proceedings of the 13th International Conference on Natural Language Processing: 2016. p. 55–64.

  11. Li F, Ji D, Wei X, Qian T. A transition-based model for jointly extracting drugs, diseases and adverse drug events. In: 2015 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE: 2015. https://doi.org/10.1109/bibm.2015.7359750.

  12. Wang A, Wang J, Lin H, Zhang J, Yang Z, Xu K. A multiple distributed representation method based on neural network for biomedical event extraction. BMC Med Informa Decis Mak. 2017; 17(3):171.

    Article  Google Scholar 

  13. Li L, Liu Y, Qin M. Extracting Biomedical Events with Parallel Multi-Pooling Convolutional Neural Networks. IEEE/ACM Trans Comput Biol Bioinforma. 2018:1. https://doi.org/10.1109/tcbb.2018.2868078.

  14. Björne J, Salakoski T. Biomedical Event Extraction Using Convolutional Neural Networks and Dependency Parsing. In: Proceedings of the BioNLP 2018 workshop. Association for Computational Linguistics: 2018. https://doi.org/10.18653/v1/w18-2311.

  15. Li L, Wan J, Zheng J, Wang J. Biomedical event extraction based on GRU integrating attention mechanism. BMC Bioinforma. 2018; 19(9):177.

    Google Scholar 

  16. Zhu Q, Li X, Conesa A, Pereira C. GRAM-CNN: a deep learning approach with local context for named entity recognition in biomedical text. Bioinformatics. 2017; 34(9):1547–54.

    Article  Google Scholar 

  17. Li L, Jiang Y. Biomedical named entity recognition based on the two channels and sentence-level reading control conditioned LSTM-CRF. In: 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE: 2017. https://doi.org/10.1109/bibm.2017.8217679.

  18. Wang Y, Wang J, Lin H, Tang X, Zhang S, Li L. Bidirectional long short-term memory with CRF for detecting biomedical event trigger in FastText semantic space. BMC Bioinforma. 2018; 19(20):507.

    Article  CAS  Google Scholar 

  19. Raj D, Sahu S, Anand A. Learning local and global contexts using a convolutional recurrent network model for relation classification in biomedical text. In: Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL 2017). Association for Computational Linguistics: 2017. https://doi.org/10.18653/v1/k17-1032.

  20. Zheng S, Hao Y, Lu D, Bao H, Xu J, Hao H, et al.Joint entity and relation extraction based on a hybrid neural network. Neurocomputing. 2017; 257:59–66.

    Article  Google Scholar 

  21. Li F, Zhang M, Fu G, Ji D. A neural joint model for entity and relation extraction from biomedical text. BMC Bioinforma. 2017; 18(1):198.

    Article  CAS  Google Scholar 

  22. Miwa M, Bansal M. End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Association for Computational Linguistics: 2016. https://doi.org/10.18653/v1/p16-1105.

  23. Kim S, Fiorini N, Wilbur WJ, Lu Z. Bridging the gap: incorporating a semantic similarity measure for effectively mapping PubMed queries to documents. J Biomed Inform. 2017; 75:122–7.

    Article  Google Scholar 

  24. Pyysalo S, Ohta T, Miwa M, Cho HC, Tsujii J, Ananiadou S. Event extraction across multiple levels of biological organization. Bioinformatics. 2012; 28(18):i575–81.

    Article  CAS  Google Scholar 

  25. Zhou D, Zhong D, He Y. Event trigger identification for biomedical events extraction using domain knowledge. Bioinformatics. 2014; 30(11):1587–94.

    Article  Google Scholar 

  26. Glorot X, Bengio Y. Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics. Springer-Verlag: 2010. p. 249–256.

  27. Kim JD, Ohta T, Pyysalo S, Kano Y, Tsujii J. Extracting bio-molecular events from literature the bionlp09 shared task. Comput Intell. 2011; 27(4):513–40.

    Article  Google Scholar 

  28. Zhou D, Zhong D. A semi-supervised learning framework for biomedical event extraction based on hidden topics. Artif Intell Med. 2015; 64(1):51–8.

    Article  Google Scholar 

  29. Box JF, et al.Guinness, Gosset, Fisher, and small samples. Stat Sci. 1987; 2(1):45–52.

    Article  Google Scholar 

  30. Ma X, Hovy E. End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Association for Computational Linguistics: 2016. https://doi.org/10.18653/v1/p16-1101.

  31. Hanson J, Yang Y, Paliwal K, Zhou Y. Improving protein disorder prediction by deep bidirectional long short-term memory recurrent neural networks. Bioinformatics. 2016; 33(5):685–92.

    Google Scholar 

  32. Ratinov L, Roth D. Design challenges and misconceptions in named entity recognition. In: Proceedings of the Thirteenth Conference on Computational Natural Language Learning - CoNLL ’09: 2009. https://doi.org/10.3115/1596374.1596399.

  33. Bengio S, Vinyals O, Jaitly N, Shazeer N. Scheduled sampling for sequence prediction with recurrent neural networks. Adv Neural Inf Process Syst. 2015:1171–9. Curran Associates, Inc.

Download references

Acknowledgements

Not applicable.

Funding

This work was supported by the National Key Technologies R&D Program [2017YFA0505502] and the Natural Science Foundation of Anhui Province [1508085MF128]. The funding body played no role in the study and collection, analysis, and interpretation of data and in writing the manuscript.

Author information

Authors and Affiliations

Authors

Contributions

ZL and ZH designed the study and drafted the manuscript. ZL conducted the experiments and ZH arranged the study plan. Both authors read and approved the final manuscript.

Corresponding author

Correspondence to Haoran Zheng.

Ethics declarations

Ethics approval and consent to participate

Not applicable

Consent for publication

Not applicable

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Additional file 1

Supplementary.pdf contains detailed structure of TR/RC/EE (Figure S1/Figure S2/Figure S3), statistics of event structure for CG/PC/MLEE (Table S1/Table S2/Table S3), performance of experiments for CG/PC/MLEE (Table S4/Table S5/Table S6), t-test result for performance of CG/PC/MLEE (Table S7/Table S8/Table S9), statistics of the ignored cases (Table S10) and detailed hyper-parameters (Table S11).

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhu, L., Zheng, H. Biomedical event extraction with a novel combination strategy based on hybrid deep neural networks. BMC Bioinformatics 21, 47 (2020). https://doi.org/10.1186/s12859-020-3376-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12859-020-3376-2

Keywords