z-logo
open-access-imgOpen Access
Improving Non-Autoregressive Machine Translation via Autoregressive Training
Author(s) -
Shuheng Wang,
Shumin Shi,
Heyan Huang,
Wei Zhang
Publication year - 2021
Publication title -
journal of physics. conference series
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.21
H-Index - 85
eISSN - 1742-6596
pISSN - 1742-6588
DOI - 10.1088/1742-6596/2031/1/012045
Subject(s) - autoregressive model , computer science , nat , task (project management) , machine translation , artificial intelligence , machine learning , encoder , translation (biology) , decoding methods , speech recognition , algorithm , computer network , engineering , operating system , econometrics , biochemistry , chemistry , systems engineering , messenger rna , gene , economics
In recent years, non-autoregressive machine translation has attracted many researchers’ attentions. Non-autoregressive translation (NAT) achieves faster decoding speed at the cost of translation accuracy compared with autoregressive translation (AT). Since NAT and AT models have similar architecture, a natural idea is to use AT task assisting NAT task. Previous works use curriculum learning or distillation to improve the performance of NAT model. However, they are complex to follow and diffucult to be integrated into some new works. So in this paper, to make it easy, we introduce a multi-task framework to improve the performance of NAT task. Specially, we use a fully shared encoder-decoder network to train NAT task and AT task simultaneously. To evaluate the performance of our model, we conduct experiments on serval benchmask tasks, including WMT14 EN-DE, WMT16 EN-RO and IWSLT14 DE-EN. The experimental results demonstrate that our model achieves improvements but still keeps simple.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here