Skip to main content Skip to main navigation

Publikation

Incremental Domain Adaptation for Neural Machine Translation in Low-Resource Settings

Marimuthu Kalimuthu; Michael Barz; Daniel Sonntag
In: Proceedings of the Fourth Arabic Natural Language Processing Workshop. Arabic Natural Language Processing Workshop (WANLP-2019), 4th, located at ACL, August 1-2, Florence, Italy, Pages 1-10, Association for Computational Linguistics, 8/2019.

Zusammenfassung

We study the problem of incremental domain adaptation of a generic neural machine translation model with limited resources (e.g., budget and time) for human translations or model training. In this paper, we propose a novel query strategy for selecting {``}unlabeled{''} samples from a new domain based on sentence embeddings for Arabic. We accelerate the fine-tuning process of the generic model to the target domain. Specifically, our approach estimates the informativeness of instances from the target domain by comparing the distance of their sentence embeddings to embeddings from the generic domain. We perform machine translation experiments (Ar-to-En direction) for comparing a random sampling baseline with our new approach, similar to active learning, using two small update sets for simulating the work of human translators. For the prescribed setting we can save more than 50{\%} of the annotation costs without loss in quality, demonstrating the effectiveness of our approach.