HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

A Character-Level Decoder without Explicit Segmentation for Neural Machine Translation

Junyoung Chung; Kyunghyun Cho; Yoshua Bengio

A Character-Level Decoder without Explicit Segmentation for Neural Machine Translation

Abstract

The existing machine translation systems, whether phrase-based or neural, have relied almost exclusively on word-level modelling with explicit segmentation. In this paper, we ask a fundamental question: can neural machine translation generate a character sequence without any explicit segmentation? To answer this question, we evaluate an attention-based encoder-decoder with a subword-level encoder and a character-level decoder on four language pairs--En-Cs, En-De, En-Ru and En-Fi-- using the parallel corpora from WMT'15. Our experiments show that the models with a character-level decoder outperform the ones with a subword-level decoder on all of the four language pairs. Furthermore, the ensembles of neural models with a character-level decoder outperform the state-of-the-art non-neural machine translation systems on En-Cs, En-De and En-Fi and perform comparably on En-Ru.

Code Repositories

nyu-dl/dl4mt-cdec
Mentioned in GitHub
nyu-dl/dl4mt-c2c
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
machine-translation-on-wmt2015-english-germanEnc-Dec Att (char)
BLEU score: 23.5
machine-translation-on-wmt2015-english-germanEnc-Dec Att (BPE)
BLEU score: 21.7

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
A Character-Level Decoder without Explicit Segmentation for Neural Machine Translation | Papers | HyperAI