Conference paper Open Access

Domain Adaptation of Document-Level NMT in IWSLT19

Popel, Martin; Federmann, Christian

We describe our four NMT systems submitted to the IWSLT19 shared task in English→Czech text-to-text translation of TED talks. The goal of this study is to understand the interactions between document-level NMT and domain adaptation. All our systems are based on the Transformer model implemented in the Tensor2Tensor framework. Two of the systems serve as baselines, which are not adapted to the TED talks domain: SENTBASE is trained on single sen- tences, DOCBASE on multi-sentence (document-level) sequences. The other two submitted systems are adapted to TED talks: SENTFINE is fine-tuned on single sentences, DOCFINE is fine-tuned on multi-sentence sequences. We present both automatic-metrics evaluation and manual analysis of the translation quality, focusing on the differences between the four systems.

Files (153.5 kB)
Name Size
153.5 kB Download
All versions This version
Views 105105
Downloads 114114
Data volume 17.5 MB17.5 MB
Unique views 9797
Unique downloads 109109


Cite as