Awesome
Test Transformer with AdaBelief Optimizer
This repo is based on fairseq
repository (lates version, 1.0.0a0) https://github.com/pytorch/fairseq <br>
An implementation of AdaBelief optimizer compatible with fairseq
is in fairseq/optim
.<br>
The original AdaBelief implementation is in https://github.com/juntang-zhuang/Adabelief-Optimizer<br>
Code for transformer to work with PyTorch 1.1 and CUDA9.0 is at: https://github.com/juntang-zhuang/transformer-adabelief<br>
Link to AdaBelief repository https://github.com/juntang-zhuang/Adabelief-Optimizer <br>
Dependencies
PyTorch==1.6.0
How to run on IWSLT14 DE-EN
Install current package
pip install --editable .
Prepare data
cd examples/translation/
bash prepare-iwslt14.sh
cd ../..
Preprocess/binarize the data
TEXT=examples/translation/iwslt14.tokenized.de-en
fairseq-preprocess --source-lang de --target-lang en \
--trainpref $TEXT/train --validpref $TEXT/valid --testpref $TEXT/test \
--destdir data-bin/iwslt14.tokenized.de-en \
--workers 20
run with Adam optimizer
sh run_adam.sh
Results saved in folder adam
run with AdaBelief optimizer
sh run_adabelief.sh
Results saved in folder adabelief
Results (BLEU score)
Adam | Adabelief |
---|---|
35.02 | 35.17 |