8000 GitHub - ictnlp/RSI-NAT: Source code for "Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation"
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content
/ RSI-NAT Public

Source code for "Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation"

License

BSD-3-Clause, BSD-3-Clause licenses found

Licenses found

BSD-3-Clause
LICENSE
BSD-3-Clause
LICENSE_nyu
Notifications You must be signed in to change notification settings

ictnlp/RSI-NAT

Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation

PyTorch implementation of the models described in the paper Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation .

Dependencies

Python

  • Python 3.6
  • PyTorch 0.4
  • Numpy
  • NLTK
  • torchtext
  • torchvision
  • revtok
  • multiset
  • ipdb

GPU

  • CUDA (we recommend using the latest version. The version 8.0 was used in all our experiments.)

Related code

Downloading Datasets

The original translation corpora can be downloaded from (IWLST'16 En-De, WMT'16 En-Ro, WMT'14 En-De). We recommend you to download the preprocessed corpora released in dl4mt-nonauto.

Before you run the code

Set correct path to data in data_path() function located in data.py:

Training New Models

Train a NAT model using the cross-entropy loss. This process usually takes about 10 days. You can download our pretrained models here

IWSLT

$ sh train_iwslt.sh

WMT14 En-De

$ sh rf_wmt.sh

Finetuning (RF-NAT)

Take a checkpoint p 8000 re-trained non-autoregressive model and finetune the checkpoint using the RF-NAT algorithm. This process usually takes about 1 days. If you want to use the origin REINFORCE, change the flag --nat_finetune to --rf_finetune.

IWSLT

$ sh rf_iwslt.sh

WMT14 En-De

$ sh rf_wmt.sh

Training the Length Prediction Model

Take a finetuned checkpoint and train the length prediction model. This process usually takes about 1 day.

IWSLT

$ sh tune_iwslt.sh

WMT14 En-De

$ sh tune_wmt.sh

Decoding

Decode the test set. This process usually takes about 20 seconds.

IWSLT

$ sh decode_iwslt.sh

WMT14 En-De

$ sh decode_wmt.sh

Citation

If you find the resources in this repository useful, please consider citing:

@inproceedings{shao2019retrieving,
    title = "Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation",
    author = "Shao, Chenze  and
      Feng, Yang  and
      Zhang, Jinchao  and
      Meng, Fandong  and
      Chen, Xilin  and
      Zhou, Jie",
    booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
    month = jul,
    year = "2019",
    url = "https://www.aclweb.org/anthology/P19-1288",
    pages = "3013--3024",
}

About

Source code for "Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation"

Resources

License

BSD-3-Clause, BSD-3-Clause licenses found

Licenses found

BSD-3-Clause
LICENSE
BSD-3-Clause
LICENSE_nyu

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published
0