Proceedings of the Third Conference on Machine Translation: Shared Task Papers 2018
DOI: 10.18653/v1/w18-6439
|View full text |Cite
|
Sign up to set email alerts
|

The MeMAD Submission to the WMT18 Multimodal Translation Task

Abstract: This paper describes the MeMAD project entry to the WMT Multimodal Machine Translation Shared Task.We propose adapting the Transformer neural machine translation (NMT) architecture to a multi-modal setting. In this paper, we also describe the preliminary experiments with textonly translation systems leading us up to this choice.We have the top scoring system for both English-to-German and English-to-French, according to the automatic metrics for flickr18.Our experiments show that the effect of the visual featu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
40
1
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 53 publications
(47 citation statements)
references
References 17 publications
(24 reference statements)
3
40
1
1
Order By: Relevance
“…The CUNI submissions use two architectures based on the self-attentive Transformer model (Vaswani et al, 2017). For German and Czech, a language model is used to extract pseudo-in-ID Participating team AFRL-OHIOSTATE Air Force Research Laboratory & Ohio State University (Gwinnup et al, 2018) CUNI Univerzita Karlova v Praze (Helcl et al, 2018) LIUMCVC Laboratoire d'Informatique de l'Université du Maine & Universitat Autonoma de Barcelona Computer Vision Center (Caglayan et al, 2018) MeMAD Aalto University, Helsinki University & EURECOM (Grönroos et al, 2018) OSU-BAIDU Oregon State University & Baidu Research (Zheng et al, 2018) SHEF University of Sheffield UMONS Université de Mons (Delbrouck and Dupont, 2018) Table 5: Participants in the WMT18 multimodal machine translation shared task.…”
Section: Cuni (Task 1)mentioning
confidence: 99%
“…The CUNI submissions use two architectures based on the self-attentive Transformer model (Vaswani et al, 2017). For German and Czech, a language model is used to extract pseudo-in-ID Participating team AFRL-OHIOSTATE Air Force Research Laboratory & Ohio State University (Gwinnup et al, 2018) CUNI Univerzita Karlova v Praze (Helcl et al, 2018) LIUMCVC Laboratoire d'Informatique de l'Université du Maine & Universitat Autonoma de Barcelona Computer Vision Center (Caglayan et al, 2018) MeMAD Aalto University, Helsinki University & EURECOM (Grönroos et al, 2018) OSU-BAIDU Oregon State University & Baidu Research (Zheng et al, 2018) SHEF University of Sheffield UMONS Université de Mons (Delbrouck and Dupont, 2018) Table 5: Participants in the WMT18 multimodal machine translation shared task.…”
Section: Cuni (Task 1)mentioning
confidence: 99%
“…Each image in this dataset is associated with up to 5 independently annotated English captions, with a total of 616,767 captions. Though originally a monolingual dataset, the dataset's large size makes it useful for data augmentation methods for image-guided translation, as demonstrated in Grönroos et al (2018). There has also been some effort to add other languages to COCO.…”
Section: Flickr8kmentioning
confidence: 99%
“…Recently, some Transformer-based multimodal NMT models have been proposed. Grönroos et al [11] added a gating layer to each output of the Transformer encoder and decoder, and their model uses visual features in the gate. They showed that the proposed gating layer in the encoder decreases ambiguity in encoding source language sentences and that in the decoder suppresses the outputs of unnecessary words.…”
Section: Previous Multimodal Nmtmentioning
confidence: 99%