HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Gumbel-Attention for Multi-modal Machine Translation

Pengbo Liu Hailong Cao Tiejun Zhao

Gumbel-Attention for Multi-modal Machine Translation

Abstract

Multi-modal machine translation (MMT) improves translation quality by introducing visual information. However, the existing MMT model ignores the problem that the image will bring information irrelevant to the text, causing much noise to the model and affecting the translation quality. This paper proposes a novel Gumbel-Attention for multi-modal machine translation, which selects the text-related parts of the image features. Specifically, different from the previous attention-based method, we first use a differentiable method to select the image information and automatically remove the useless parts of the image features. Experiments prove that our method retains the image features related to the text, and the remaining parts help the MMT model generates better translations.

Benchmarks

BenchmarkMethodologyMetrics
multimodal-machine-translation-on-multi30kGumbel-Attention MMT
BLEU (EN-DE): 39.2
Meteor (EN-DE): 57.8

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Gumbel-Attention for Multi-modal Machine Translation | Papers | HyperAI