A2R2: Robust Unsupervised Neural Machine Translation With Adversarial Attack and Regularization on Representations

Unsupervised neural machine translation (UNMT) has recently achieved significant progress without requirement on any parallel data. The models for UNMT are typically the sequence-to-sequence architecture with an encoder to map sentences in different languages to a shared latent space, and a decoder...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2021, Vol.9, p.19990-19998
Hauptverfasser: Yu, Heng, Luo, Haoran, Yi, Yuqi, Cheng, Fan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Unsupervised neural machine translation (UNMT) has recently achieved significant progress without requirement on any parallel data. The models for UNMT are typically the sequence-to-sequence architecture with an encoder to map sentences in different languages to a shared latent space, and a decoder to generate their corresponding translation. Denoising autoencoding and back-translation are called in every iteration for the models to learn the relationship of sentence pairs in languages or between languages. However, sentences generated by the noise model of autoencoding or the reverse model of back-translation are normally different from those written by humans, which may cause inference bias. In this paper, we propose a regularization method for back-translation to explicitly draw representations of sentence pairs closer in the shared space. To enhance the robustness to sentences after autoencoding or back-translation, the adversarial attack on representations is applied. Experiments on unsupervised English \leftrightarrow French, English \leftrightarrow German and English \leftrightarrow Romanian benchmarks show that our approach outperforms the cross-lingual language model (XLM) baseline by 0.4\sim 1.8 BLEU scores. Additionally, the boost on noisy test sets in most translation directions is over 5 BLEU scores.
ISSN:2169-3536
2169-3536
DOI:10.1109/ACCESS.2021.3054935