Iterative training of unsupervised neural and statistical machine translation systems
ACM Transactions on Asian and Low-Resource Language Information Processing …, 2020•dl.acm.org
Recent work achieved remarkable results in training neural machine translation (NMT)
systems in a fully unsupervised way, with new and dedicated architectures that only rely on
monolingual corpora. However, previous work also showed that unsupervised statistical
machine translation (USMT) performs better than unsupervised NMT (UNMT), especially for
distant language pairs. To take advantage of the superiority of USMT over UNMT, and
considering that SMT suffers from well-known limitations overcome by NMT, we propose to …
systems in a fully unsupervised way, with new and dedicated architectures that only rely on
monolingual corpora. However, previous work also showed that unsupervised statistical
machine translation (USMT) performs better than unsupervised NMT (UNMT), especially for
distant language pairs. To take advantage of the superiority of USMT over UNMT, and
considering that SMT suffers from well-known limitations overcome by NMT, we propose to …
Recent work achieved remarkable results in training neural machine translation (NMT) systems in a fully unsupervised way, with new and dedicated architectures that only rely on monolingual corpora. However, previous work also showed that unsupervised statistical machine translation (USMT) performs better than unsupervised NMT (UNMT), especially for distant language pairs. To take advantage of the superiority of USMT over UNMT, and considering that SMT suffers from well-known limitations overcome by NMT, we propose to define UNMT as NMT trained with the supervision of synthetic parallel data generated by USMT. This way we can exploit USMT up to its limits while ultimately relying on full-fledged NMT models to generate translations. We show significant improvements in translation quality over previous work and also that further improvements can be obtained by alternatively and iteratively training USMT and UNMT. Without the need of a dedicated architecture for UNMT, our simple approach can straightforwardly benefit from any recent and future advances in supervised NMT. Our systems achieve a new state-of-the-art for unsupervised machine translation in all of our six translation tasks for five diverse language pairs, surpassing even supervised SMT or NMT in some tasks. Furthermore, our analysis shows how crucial the comparability between the monolingual corpora used for unsupervised training is in improving translation quality.
ACM Digital Library
Showing the best result for this search. See all results