HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

SenseBERT: Driving Some Sense into BERT

Yoav Levine; Barak Lenz; Or Dagan; Ori Ram; Dan Padnos; Or Sharir; Shai Shalev-Shwartz; Amnon Shashua; Yoav Shoham

SenseBERT: Driving Some Sense into BERT

Abstract

The ability to learn from large unlabeled corpora has allowed neural language models to advance the frontier in natural language understanding. However, existing self-supervision techniques operate at the word form level, which serves as a surrogate for the underlying semantic content. This paper proposes a method to employ weak-supervision directly at the word sense level. Our model, named SenseBERT, is pre-trained to predict not only the masked words but also their WordNet supersenses. Accordingly, we attain a lexical-semantic level language model, without the use of human annotation. SenseBERT achieves significantly improved lexical understanding, as we demonstrate by experimenting on SemEval Word Sense Disambiguation, and by attaining a state of the art result on the Word in Context task.

Benchmarks

BenchmarkMethodologyMetrics
natural-language-inference-on-qnliSenseBERT-base 110M
Accuracy: 90.6%
natural-language-inference-on-rteSenseBERT-base 110M
Accuracy: 67.5%
word-sense-disambiguation-on-words-in-contextSenseBERT-large 340M
Accuracy: 72.1
word-sense-disambiguation-on-words-in-contextSenseBERT-base 110M
Accuracy: 70.3

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
SenseBERT: Driving Some Sense into BERT | Papers | HyperAI