Published September 1, 2020 | Version v1
Conference paper Open

FinEst BERT and CroSloEngual BERT: less is more in multilingual models

  • 1. University of Ljubljana, Ljubljana, Slovenia

Description

Large pre-trained masked language models have become state-of-the-art solutions for many NLP problems. The research has been mostly focused on English language, though. While massively multilingual models exist, studies have shown that monolingual models produce much better results. We train two trilingual BERT-like models, one for Finnish, Estonian, and English, the other for Croatian, Slovenian, and English. We evaluate their performance on several downstream tasks, NER, POS-tagging, and dependency parsing, using the multilingual BERT and XLM-R as baselines. The newly created FinEst BERT and CroSloEngual BERT improve the results on all tasks in most monolingual and cross-lingual situations.

Files

Ulčar-Robnik-Šikonja2020_Chapter_FinEstBERTAndCroSloEngualBERT.pdf

Files (237.4 kB)

Additional details

Funding

EMBEDDIA – Cross-Lingual Embeddings for Less-Represented Languages in European News Media 825153
European Commission