Model | Micro-averaged | Macro-averaged | ||||
---|---|---|---|---|---|---|
P | R | F1 | P | R | F1 | |
LSTM | 0.486 | 0.388 | 0.431 | 0.537 | 0.443 | 0.395 |
+ Self-attention | 0.490 | 0.391 | 0.435 | 0.550 | 0.452 | 0.405 |
+ Knowledge-based attention | 0.495 | 0.395 | 0.440 | 0.559 | 0.465 | 0.417 |
GRU | 0.484 | 0.387 | 0.430 | 0.548 | 0.454 | 0.406 |
+ Self-attention | 0.491 | 0.393 | 0.436 | 0.558 | 0.461 | 0.414 |
+ Knowledge-based attention | 0.495 | 0.395 | 0.439 | 0.551 | 0.454 | 0.407 |
BiLSTM | 0.486 | 0.389 | 0.432 | 0.541 | 0.446 | 0.398 |
+ Self-attention | 0.493 | 0.394 | 0.438 | 0.552 | 0.456 | 0.408 |
+ Knowledge-based attention | 0.499 | 0.400 | 0.444 | 0.559 | 0.461 | 0.414 |
BiGRU | 0.487 | 0.389 | 0.433 | 0.542 | 0.446 | 0.398 |
+ Self-attention | 0.497 | 0.397 | 0.441 | 0.558 | 0.462 | 0.415 |
+ Knowledge-based attention | 0.501 | 0.400 | 0.445 | 0.562 | 0.464 | 0.416 |
Hierarchical-ConvNet | 0.468 | 0.375 | 0.416 | 0.522 | 0.429 | 0.381 |
+ Self-attention | 0.473 | 0.378 | 0.420 | 0.529 | 0.434 | 0.386 |
+ Knowledge-based attention | 0.483 | 0.386 | 0.429 | 0.532 | 0.441 | 0.392 |