Published July 3, 2022 | Version v1
Conference paper Open

Non-Autoregressive Machine Translation: It's Not as Fast as it Seems

  • 1. University of Edinburgh

Description

Efficient machine translation models are com- mercially important as they can increase infer- ence speeds, and reduce costs and carbon emis- sions. Recently, there has been much interest in non-autoregressive (NAR) models, which promise faster translation. In parallel to the research on NAR models, there have been suc- cessful attempts to create optimized autore- gressive models as part of the WMT shared task on efficient translation. In this paper, we point out flaws in the evaluation methodology present in the literature on NAR models and we provide a fair comparison between a state- of-the-art NAR model and the autoregressive submissions to the shared task. We make the case for consistent evaluation of NAR models, and also for the importance of comparing NAR models with other widely used methods for im- proving efficiency. We run experiments with a connectionist-temporal-classification-based (CTC) NAR model implemented in C++ and compare it with AR models using wall clock times. Our results show that, although NAR models are faster on GPUs, with small batch sizes, they are almost always slower under more realistic usage conditions. We call for more realistic and extensive evaluation of NAR models in future work.

Files

Non_Autoregressive_Machine_Translation__It_s_Not_as_Fast_as_it_Seems.pdf

Files (235.7 kB)

Additional details

Funding

GoURMET – Global Under-Resourced MEedia Translation 825299
European Commission