HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Multimodal Machine Translation through Visuals and Speech

Umut Sulubacak Ozan Caglayan Stig-Arne Grönroos Aku Rouhe Desmond Elliott Lucia Specia Jörg Tiedemann

Multimodal Machine Translation through Visuals and Speech

Abstract

Multimodal machine translation involves drawing information from more than one modality, based on the assumption that the additional modalities will contain useful alternative views of the input data. The most prominent tasks in this area are spoken language translation, image-guided translation, and video-guided translation, which exploit audio and visual modalities, respectively. These tasks are distinguished from their monolingual counterparts of speech recognition, image captioning, and video captioning by the requirement of models to generate outputs in a different language. This survey reviews the major data resources for these tasks, the evaluation campaigns concentrated around them, the state of the art in end-to-end and pipeline approaches, and also the challenges in performance evaluation. The paper concludes with a discussion of directions for future research in these areas: the need for more expansive and challenging datasets, for targeted evaluations of model performance, and for multimodality in both the input and output space.

Benchmarks

BenchmarkMethodologyMetrics
multimodal-machine-translation-on-multi30kCaglayan
BLEU (EN-DE): 39.4
Meteor (EN-DE): 58.7

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp