Published October 4, 2021 | Version v1
Conference paper Open

An evaluation of BERT and Doc2Vec model on the IPTC Subject Codes prediction dataset

  • 1. TRIKODER DOO, Zagreb
  • 2. University of Ljubljana, Ljubljana, Slovenia
  • 3. Jožef Stefan Institute

Description

Large pretrained language models like BERT have shown excellent generalization properties and have advanced the state of the art on various NLP tasks. In this paper we evaluate Finnish BERT (FinBERT) model on the IPTC Subject Codes prediction task. We compare it to a simpler Doc2Vec model used as a baseline. Due to hierarchical nature of IPTC Subject Codes, we also evaluate the effect of encoding the hierarchy in the network layer topology. Contrary to our expectations, a simpler baseline Doc2Vec model clearly outperforms the more complex FinBERT model and our attempts to encode hierarchy in a prediction network do not yield systematic improvement.

Files

Pranjicetal.pdf

Files (422.6 kB)

Name Size Download all
md5:7345af9f6c4af4ac18e49ebc9aa7b3c0
422.6 kB Preview Download

Additional details

Funding

European Commission
EMBEDDIA - Cross-Lingual Embeddings for Less-Represented Languages in European News Media 825153