HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

VarMAE: Pre-training of Variational Masked Autoencoder for Domain-adaptive Language Understanding

Dou Hu Xiaolong Hou Xiyang Du Mengyuan Zhou Lianxin Jiang Yang Mo Xiaofeng Shi

VarMAE: Pre-training of Variational Masked Autoencoder for Domain-adaptive Language Understanding

Abstract

Pre-trained language models have achieved promising performance on general benchmarks, but underperform when migrated to a specific domain. Recent works perform pre-training from scratch or continual pre-training on domain corpora. However, in many specific domains, the limited corpus can hardly support obtaining precise representations. To address this issue, we propose a novel Transformer-based language model named VarMAE for domain-adaptive language understanding. Under the masked autoencoding objective, we design a context uncertainty learning module to encode the token's context into a smooth latent distribution. The module can produce diverse and well-formed contextual representations. Experiments on science- and finance-domain NLU tasks demonstrate that VarMAE can be efficiently adapted to new domains with limited resources.

Benchmarks

BenchmarkMethodologyMetrics
citation-intent-classification-on-acl-arcVarMAE
Macro-F1: Not reported
Micro-F1: 76.50
participant-intervention-comparison-outcomeVarMAE
F1: 76.01

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
VarMAE: Pre-training of Variational Masked Autoencoder for Domain-adaptive Language Understanding | Papers | HyperAI