Command Palette
Search for a command to run...
Stefan Schweter Alan Akbik

Abstract
Current state-of-the-art approaches for named entity recognition (NER) typically consider text at the sentence-level and thus do not model information that crosses sentence boundaries. However, the use of transformer-based models for NER offers natural options for capturing document-level features. In this paper, we perform a comparative evaluation of document-level features in the two standard NER architectures commonly considered in the literature, namely "fine-tuning" and "feature-based LSTM-CRF". We evaluate different hyperparameters for document-level features such as context window size and enforcing document-locality. We present experiments from which we derive recommendations for how to model document context and present new state-of-the-art scores on several CoNLL-03 benchmark datasets. Our approach is integrated into the Flair framework to facilitate reproduction of our experiments.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| named-entity-recognition-ner-on-conll-2003 | FLERT XLM-R | F1: 94.09 |
| named-entity-recognition-on-conll-2002 | FLERT XLM-R | F1: 90.14 |
| named-entity-recognition-on-conll-2002-dutch | FLERT XLM-R | F1: 95.21 |
| named-entity-recognition-on-conll-2003-german | FLERT XLM-R | F1: 88.34 |
| named-entity-recognition-on-conll-2003-german-1 | FLERT XLM-R | F1: 92.23 |
| named-entity-recognition-on-findvehicle | FLERT | F1 Score: 80.9 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.