HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Text-to-Text Pre-Training for Data-to-Text Tasks

Mihir Kale Abhinav Rastogi

Text-to-Text Pre-Training for Data-to-Text Tasks

Abstract

We study the pre-train + fine-tune strategy for data-to-text tasks. Our experiments indicate that text-to-text pre-training in the form of T5, enables simple, end-to-end transformer based models to outperform pipelined neural architectures tailored for data-to-text generation, as well as alternative language model based pre-training techniques such as BERT and GPT-2. Importantly, T5 pre-training leads to better generalization, as evidenced by large improvements on out-of-domain test sets. We hope our work serves as a useful baseline for future research, as transfer learning becomes ever more prevalent for data-to-text tasks.

Code Repositories

google-research-datasets/ToTTo
Official
Mentioned in GitHub
shark-nlp/cont
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
data-to-text-generation-on-multiwoz-2-1T5-Base
BLEU: 35.1
data-to-text-generation-on-tottoT5-3B
BLEU: 49.5
PARENT: 58.4
data-to-text-generation-on-webnlgT5-Base
BLEU: 64.7
data-to-text-generation-on-webnlg-full-1T5-Large
BLEU: 57.1

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Text-to-Text Pre-Training for Data-to-Text Tasks | Papers | HyperAI