HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

MeDAL: Medical Abbreviation Disambiguation Dataset for Natural Language Understanding Pretraining

Zhi Wen Xing Han Lu Siva Reddy

MeDAL: Medical Abbreviation Disambiguation Dataset for Natural Language Understanding Pretraining

Abstract

One of the biggest challenges that prohibit the use of many current NLP methods in clinical settings is the availability of public datasets. In this work, we present MeDAL, a large medical text dataset curated for abbreviation disambiguation, designed for natural language understanding pre-training in the medical domain. We pre-trained several models of common architectures on this dataset and empirically showed that such pre-training leads to improved performance and convergence speed when fine-tuning on downstream medical tasks.

Code Repositories

mcGill-NLP/medal
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
mortality-prediction-on-mimic-iiiLSTM+SA (pretrained)
Accuracy: 0.8298
mortality-prediction-on-mimic-iiiELECTRA (pretrained)
Accuracy: 0.8443
mortality-prediction-on-mimic-iiiELECTRA (from scratch)
Accuracy: 0.8325
mortality-prediction-on-mimic-iiiLSTM (pretrained)
Accuracy: 0.828
mortality-prediction-on-mimic-iiiLSTM+SA (from scratch)
Accuracy: 0.7996

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
MeDAL: Medical Abbreviation Disambiguation Dataset for Natural Language Understanding Pretraining | Papers | HyperAI