Understanding and Improving Morphological Learning in the Neural Machine Translation Decoder

Fahim Dalvi, Nadir Durrani, Hassan Sajjad, Yonatan Belinkov, Stephan Vogel

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

End-to-end training makes the neural machine translation (NMT) architecture simpler, yet elegant compared to traditional statistical machine translation (SMT). However, little is known about linguistic patterns of morphology, syntax and semantics learned during the training of NMT systems, and more importantly, which parts of the architecture are responsible for learning each of these phenomenon. In this paper we i) analyze how much morphology an NMT decoder learns, and ii) investigate whether injecting target morphology in the decoder helps it to produce better translations. To this end we present three methods: i) simultaneous translation, ii) joint-data learning, and iii) multi-task learning. Our results show that explicit morphological information helps the decoder learn target language morphology and improves the translation quality by 0.2--0.6 BLEU points.
Original languageUndefined/Unknown
Title of host publicationProceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers)
Place of PublicationTaipei, Taiwan
PublisherAsian Federation of Natural Language Processing
Pages142-151
Number of pages10
StatePublished - 1 Nov 2017

Cite this