Command Palette
Search for a command to run...
Mihir Kale Abhinav Rastogi

Abstract
We study the pre-train + fine-tune strategy for data-to-text tasks. Our experiments indicate that text-to-text pre-training in the form of T5, enables simple, end-to-end transformer based models to outperform pipelined neural architectures tailored for data-to-text generation, as well as alternative language model based pre-training techniques such as BERT and GPT-2. Importantly, T5 pre-training leads to better generalization, as evidenced by large improvements on out-of-domain test sets. We hope our work serves as a useful baseline for future research, as transfer learning becomes ever more prevalent for data-to-text tasks.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| data-to-text-generation-on-multiwoz-2-1 | T5-Base | BLEU: 35.1 |
| data-to-text-generation-on-totto | T5-3B | BLEU: 49.5 PARENT: 58.4 |
| data-to-text-generation-on-webnlg | T5-Base | BLEU: 64.7 |
| data-to-text-generation-on-webnlg-full-1 | T5-Large | BLEU: 57.1 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.