HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

To Err Is Human, but Llamas Can Learn It Too

Agnes Luhtaru Taido Purason Martin Vainikko Maksym Del Mark Fishel

To Err Is Human, but Llamas Can Learn It Too

Abstract

This study explores enhancing grammatical error correction (GEC) through artificial error generation (AEG) using language models (LMs). Specifically, we fine-tune Llama 2-based LMs for error generation and find that this approach yields synthetic errors akin to human errors. Next, we train GEC Llama models with the help of these artificial errors and outperform previous state-of-the-art error correction models, with gains ranging between 0.8 and 6 F0.5 points across all tested languages (German, Ukrainian, and Estonian). Moreover, we demonstrate that generating errors by fine-tuning smaller sequence-to-sequence models and prompting large commercial LMs (GPT-3.5 and GPT-4) also results in synthetic errors beneficially affecting error generation models.

Code Repositories

TartuNLP/gec-llm
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
grammatical-error-correction-on-estgec-l2Llama + 1M BT + gold
F0.5: 69.97
grammatical-error-correction-on-falko-merlinLlama + 1M BT + gold
F0.5: 76.75
grammatical-error-correction-on-ua-gecLlama + 1M BT + gold
F0.5: 74.09

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
To Err Is Human, but Llamas Can Learn It Too | Papers | HyperAI