Command Palette
Search for a command to run...
Marek Rei

Abstract
We propose a sequence labeling framework with a secondary training objective, learning to predict surrounding words for every word in the dataset. This language modeling objective incentivises the system to learn general-purpose patterns of semantic and syntactic composition, which are also useful for improving accuracy on different sequence labeling tasks. The architecture was evaluated on a range of datasets, covering the tasks of error detection in learner texts, named entity recognition, chunking and POS-tagging. The novel language modeling objective provided consistent performance improvements on every benchmark, without requiring any additional annotated or unannotated data.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| grammatical-error-detection-on-conll-2014-a1 | Bi-LSTM + LMcost (trained on FCE) | F0.5: 17.86 |
| grammatical-error-detection-on-conll-2014-a2 | Bi-LSTM + LMcost (trained on FCE) | F0.5: 25.88 |
| grammatical-error-detection-on-fce | Bi-LSTM + LMcost | F0.5: 48.48 |
| part-of-speech-tagging-on-penn-treebank | Bi-LSTM + LMcost | Accuracy: 97.43 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.