Published June 1, 2022 | Version v1
Journal article Open

Sequence-to-sequence neural machine translation for English-Malay

  • 1. UOW Malaysia KDU University College

Description

Machine translation aims to translate text from a specific language into another language using computer software. In this work, we performed neural machine translation with attention implementation on English-Malay parallel corpus. We attempt to improve the model performance by rectified linear unit (ReLU) attention alignment. Different sequence-to-sequence models were trained. These models include long-short term memory (LSTM), gated recurrent unit (GRU), bidirectional LSTM (Bi-LSTM) and bidirectional GRU (Bi-GRU). In the experiment, both bidirectional models, Bi-LSTM and Bi-GRU yield a converge of below 30 epochs. Our study shows that the ReLU attention alignment improves the bilingual evaluation understudy (BLEU) translation score between score 0.26 and 1.12 across all the models as compare to the original Tanh models.

Files

27 20769.pdf

Files (324.7 kB)

Name Size Download all
md5:c42fb53de46125ddacd26cde6cfacb79
324.7 kB Preview Download