Awesome
ALFA - Meta-Learning with Adaptive Hyperparameters
Sungyong Baik, Myungsub Choi, Janghoon Choi, Heewon Kim, Kyoung Mu Lee
Source code for NeurIPS 2020 paper "Meta-Learning with Adaptive Hyperparameters" (previously titled "Adaptive Learning for Fast Adaptation")
This repository is the implementation of ALFA. The code is based off the public code of MAML++, where their reimplementation of MAML is used as the baseline.
Requirements
- Ubuntu 18.04
- Anaconda3
- Python==3.6
- PyTorch==1.5
- numpy==1.19.1
To install requirements, first download Anaconda3 and then run the following:
bash install.sh
Hardware Requirements
- GPU with memory more than 27GB for a single-GPU ResNet12 backbone second-order training.
- The current version does not support a multi-GPU setting. While running on a multi-GPU will not give errors, it will give incorrect results (due to uneven distribution of labels across GPUs).
Datasets
For miniIamgenet, the dataset can be downloaded from the link provided from MAML++ public code. make a directory named 'datasets' and place the downloaded miniImagnet under the 'datasets' directory.
Training
To train the model(s) in the paper, run this command in experiment_scripts folder:
For single GPU
bash alfa+maml.sh 0
where 0 represent GPU_ID.
Evaluation
After training is finished, the same command is run to evaluate:
For single GPU:
bash alfa+maml.sh 0