IntroVNMT: An Introspective Model for Variational Neural Machine Translation

  • Xin Sheng University of Science and Technology of China
  • Linli Xu University of Science and Technology of China
  • Junliang Guo University of Science and Technology of China
  • Jingchang Liu Hong Kong University of Science and Technology
  • Ruoyu Zhao University of Science and Technology of China
  • Yinlong Xu University of Science and Technology of China

Abstract

We propose a novel introspective model for variational neural machine translation (IntroVNMT) in this paper, inspired by the recent successful application of introspective variational autoencoder (IntroVAE) in high quality image synthesis. Different from the vanilla variational NMT model, IntroVNMT is capable of improving itself introspectively by evaluating the quality of the generated target sentences according to the high-level latent variables of the real and generated target sentences. As a consequence of introspective training, the proposed model is able to discriminate between the generated and real sentences of the target language via the latent variables generated by the encoder of the model. In this way, IntroVNMT is able to generate more realistic target sentences in practice. In the meantime, IntroVNMT inherits the advantages of the variational autoencoders (VAEs), and the model training process is more stable than the generative adversarial network (GAN) based models. Experimental results on different translation tasks demonstrate that the proposed model can achieve significant improvements over the vanilla variational NMT model.

Published
2020-04-03
Section
AAAI Technical Track: Natural Language Processing