Home

Awesome

Meta-SGD in pytorch

The only difference compared to MAML is to parametrize task learning rate in vector form when meta-training. As the authors said, we could see fast convergence and higher performance than naive MAML. For our version of Meta-SGD, we did not use other tricks to improve performances such as regularization or 1-shot meta-training. Our pytorch version of MAML is at https://github.com/jik0730/MAML-in-pytorch.

Performance comparisions to MAML

The reported performance are refered to the ones in Meta-SGD paper.

Omniglot5-way 1-shot5-way 5-shot20-way 1-shot20-way 5-shot
MAML98.7%99.9%95.8%98.9%
Ours MAML99.4%99.9%92.8%-
Meta-SGD99.5%99.9%95.9%99.0%
Ours Meta-SGD99.3%99.8%95.4%97.8%
miniImageNet5-way 1-shot5-way 5-shot20-way 1-shot20-way 5-shot
MAML48.7%63.1%16.5%19.3%
Ours MAML48.4%64.8%--
Meta-SGD50.5%64.0%17.6%28.9%
Ours Meta-SGD49.1%66.0%17.0%29.9%

TODO