MetaMT, a Meta Learning Method Leveraging Multiple Domain Data for Low Resource Machine Translation

Authors

  • Rumeng Li Umass Amherst
  • Xun Wang Umass Lowell
  • Hong Yu Umass Lowell

DOI:

https://doi.org/10.1609/aaai.v34i05.6339

Abstract

Neural machine translation (NMT) models have achieved state-of-the-art translation quality with a large quantity of parallel corpora available. However, their performance suffers significantly when it comes to domain-specific translations, in which training data are usually scarce. In this paper, we present a novel NMT model with a new word embedding transition technique for fast domain adaption. We propose to split parameters in the model into two groups: model parameters and meta parameters. The former are used to model the translation while the latter are used to adjust the representational space to generalize the model to different domains. We mimic the domain adaptation of the machine translation model to low-resource domains using multiple translation tasks on different domains. A new training strategy based on meta-learning is developed along with the proposed model to update the model parameters and meta parameters alternately. Experiments on datasets of different domains showed substantial improvements of NMT performances on a limited amount of data.

Downloads

Published

2020-04-03

How to Cite

Li, R., Wang, X., & Yu, H. (2020). MetaMT, a Meta Learning Method Leveraging Multiple Domain Data for Low Resource Machine Translation. Proceedings of the AAAI Conference on Artificial Intelligence, 34(05), 8245-8252. https://doi.org/10.1609/aaai.v34i05.6339

Issue

Section

AAAI Technical Track: Natural Language Processing