HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Does BERT Make Any Sense? Interpretable Word Sense Disambiguation with Contextualized Embeddings

Gregor Wiedemann Steffen Remus Avi Chawla Chris Biemann

Does BERT Make Any Sense? Interpretable Word Sense Disambiguation with Contextualized Embeddings

Abstract

Contextualized word embeddings (CWE) such as provided by ELMo (Peters et al., 2018), Flair NLP (Akbik et al., 2018), or BERT (Devlin et al., 2019) are a major recent innovation in NLP. CWEs provide semantic vector representations of words depending on their respective context. Their advantage over static word embeddings has been shown for a number of tasks, such as text classification, sequence tagging, or machine translation. Since vectors of the same word type can vary depending on the respective context, they implicitly provide a model for word sense disambiguation (WSD). We introduce a simple but effective approach to WSD using a nearest neighbor classification on CWEs. We compare the performance of different CWE models for the task and can report improvements above the current state of the art for two standard WSD benchmark datasets. We further show that the pre-trained BERT model is able to place polysemic words into distinct 'sense' regions of the embedding space, while ELMo and Flair NLP do not seem to possess this ability.

Code Repositories

uhh-lt/bert-sense
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
word-sense-disambiguation-on-semeval-2007kNN-BERT
F1: 60.94
word-sense-disambiguation-on-semeval-2007kNN-BERT + POS (training corpus: SemCor)
F1: 63.17
word-sense-disambiguation-on-semeval-2007-1kNN-BERT
F1: 81.20
word-sense-disambiguation-on-semeval-2007-1kNN-BERT + POS (training corpus: WNGT)
F1: 85.32
word-sense-disambiguation-on-senseval-2-1kNN-BERT
F1: 76.52
word-sense-disambiguation-on-senseval-3kNN-BERT
F1: 80.12

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Does BERT Make Any Sense? Interpretable Word Sense Disambiguation with Contextualized Embeddings | Papers | HyperAI