Command Palette
Search for a command to run...
{ro Claudio Delli Bovi Roberto Navigli Aless Raganato}

Abstract
Word Sense Disambiguation models exist in many flavors. Even though supervised ones tend to perform best in terms of accuracy, they often lose ground to more flexible knowledge-based solutions, which do not require training by a word expert for every disambiguation target. To bridge this gap we adopt a different perspective and rely on sequence learning to frame the disambiguation problem: we propose and study in depth a series of end-to-end neural architectures directly tailored to the task, from bidirectional Long Short-Term Memory to encoder-decoder models. Our extensive evaluation over standard benchmarks and in multiple languages shows that sequence learning enables more versatile all-words models that consistently lead to state-of-the-art results, even against word experts with engineered features.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| word-sense-disambiguation-on-supervised | Bi-LSTM<sub>att+LEX</sub> | SemEval 2007: 63.7* SemEval 2013: 66.4 SemEval 2015: 72.4 Senseval 2: 72.0 Senseval 3: 69.4 |
| word-sense-disambiguation-on-supervised | Bi-LSTM<sub>att+LEX+POS</sub> | SemEval 2007: 64.8* SemEval 2013: 66.9 SemEval 2015: 71.5 Senseval 2: 72.0 Senseval 3: 69.1 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.