Published July 4, 2017 | Version v1
Preprint Open

Multilingual Hierarchical Attention Networks for Document Classification

  • 1. Idiap Research Institute

Description

Hierarchical attention networks have recently achieved remarkable performance for document classification in a given language. However, when multilingual document collections are considered, training such models separately for each language entails linear parameter growth and lack of cross-language transfer. Learning a single multilingual model with fewer parameters is therefore a challenging but potentially beneficial objective. To this end, we propose multilingual hierarchical attention networks for learning document structures, with shared encoders and/or attention mechanisms across languages, using multi-task learning and an aligned semantic space as input. We evaluate the proposed models on multilingual document classification with disjoint label sets, on a large dataset which we provide, with 600k news documents in 8 languages, and 5k labels. The multilingual models outperform strong monolingual ones in low-resource as well as full-resource settings, and use fewer parameters, thus confirming their computational efficiency and the utility of cross-language transfer.

Files

multi_doc.pdf

Files (2.7 MB)

Name Size Download all
md5:d4ad12542d06dc82a08c28abef3d5cba
2.7 MB Preview Download

Additional details

Funding

SUMMA – Scalable Understanding of Multilingual Media 688139
European Commission