A Comparison of Architectures and Pretraining Methods for Contextualized Multilingual Word Embeddings

Authors

  • Niels van der Heijden University of Amsterdam
  • Samira Abnar University of Amsterdam
  • Ekaterina Shutova University of Amsterdam

DOI:

https://doi.org/10.1609/aaai.v34i05.6443

Abstract

The lack of annotated data in many languages is a well-known challenge within the field of multilingual natural language processing (NLP). Therefore, many recent studies focus on zero-shot transfer learning and joint training across languages to overcome data scarcity for low-resource languages. In this work we (i) perform a comprehensive comparison of state-of-the-art multilingual word and sentence encoders on the tasks of named entity recognition (NER) and part of speech (POS) tagging; and (ii) propose a new method for creating multilingual contextualized word embeddings, compare it to multiple baselines and show that it performs at or above state-of-the-art level in zero-shot transfer settings. Finally, we show that our method allows for better knowledge sharing across languages in a joint training setting.

Downloads

Published

2020-04-03

How to Cite

van der Heijden, N., Abnar, S., & Shutova, E. (2020). A Comparison of Architectures and Pretraining Methods for Contextualized Multilingual Word Embeddings. Proceedings of the AAAI Conference on Artificial Intelligence, 34(05), 9090-9097. https://doi.org/10.1609/aaai.v34i05.6443

Issue

Section

AAAI Technical Track: Natural Language Processing