Published November 2, 2019 | Version v1
Conference paper Open

Lexical Micro-adaptation for Neural Machine Translation

  • 1. SYSTRAN, 5 rue Feydeau, 75002 Paris (France)

Description

This work is inspired by a typical machine translation industry scenario in which translators make use of in-domain data for facilitating translation of similar or repeating sentences. We introduce a generic framework applied at inference in which a subset of segment pairs are first extracted from training data according to their similarity to the input sentences. These segments are then used to dynamically update the parameters of a generic NMT network, thus performing a lexical micro-adaptation. Our approach demonstrates strong adaptation performance to new and existing datasets including pseudo in-domain data. We evaluate our approach on a heterogeneous English-French training dataset showing accuracy gains on all evaluated domains when compared to strong adaptation baselines.

Files

IWSLT2019_paper_9.pdf

Files (432.3 kB)

Name Size Download all
md5:b145694f9a8fd725c74837681d966e22
432.3 kB Preview Download