Home

Awesome

neuralmt

A Neural Machine Translation framework for training large-scale networks on multiple nodes with multiple GPUs.

Requirements

An example for WMT15 translation task

  1. Clone neuralmt
git clone https://github.com/zomux/neuralmt
export PYTHONPATH="$PYTHONPATH:/path/to/neuralmt"
  1. Create a directory for WMT data
export WMT_ROOT="/path/to/your_wmt_folder"
mkdir $WMT_ROOT/text
mkdir $WMT_ROOT/models
  1. Tokenize de-en training corpus, and rename them to following filenames
  1. Build training data
cd /path/to/neuralmt
python examples/gru_search/preprocess.py
  1. Train on 3 GPUs
python -m deepy.multigpu.launch examples/gru_search/train.py gpu0 gpu1 gpu2
  1. Wait for several days

  2. Test your model

python examples/gru_search/test.py

(The test script only translate one sample sentence, you can modify it to translate a text file)

Note

Training on multiple machine is still in development.

Although the current framework for parallelism shall be extended to multiple machine easily, it require some works.

Some Results

Raphael Shu, 2016