Published November 7, 2021 | Version v1
Conference paper Open

Cross-lingual Sentence Embedding using Multi-Task Learning

  • 1. National University of Ireland Galway
  • 2. Huawei

Description

The scarcity of labeled training data across many languages is a significant roadblock for multilingual neural language processing. We approach the lack of in-language training data using sentence embeddings that map text written in different lan- guages, but with similar meanings, to nearby embedding space representations. The representations are produced using a dual-encoder based model trained to maximize the representational similarity between sentence pairs drawn from parallel data. The representations are enhanced using multitask training and unsupervised monolingual corpora. The effectiveness of our multilingual sentence embeddings are assessed on a comprehensive collection of monolingual, cross-lingual, and zero- shot/few-shot learning tasks.

Files

goswami2021crosslingual.pdf

Files (347.1 kB)

Name Size Download all
md5:6a491bc3896ea61b110e9670237a29c8
347.1 kB Preview Download