Fig. 5From: Investigation of improving the pre-training and fine-tuning of BERT model for biomedical relation extractionModel architectures after incorporating all outputs from the last layer. In a we show both LSTM (only black in the RNN box) and biLSTM (both black and grey line in the RNN box). a RNN on the last layer. b Attention on the last layerBack to article page