Awesome
google_landmark_retrieval_1st_place_solution
This is the 1st place solution code for google landmark retrieval 2020.
Competition link (https://www.kaggle.com/c/landmark-retrieval-2020)
Kaggle Discussion post link (https://www.kaggle.com/c/landmark-retrieval-2020/discussion/176037)
Detailed solution description arxiv link (https://arxiv.org/abs/2009.05132)
HARDWARE ENVIRONMENT : Colab TPUs
DATA : Google Landmarks Dataset v2(https://github.com/cvdfoundation/google-landmark). TF Records were used for faster training.
TRAINING(./notebooks)
- Run 'v2clean_train.ipynb' until validation loss converge. Congifuration: IMAGE_SIZE=[512,512], EFF_VER=7, BATCH_SIZE=8*8=64
- Take efficientnet backbone from step 2, and run 'v2total_train.ipynb' until validation loss converge. Configuration : NOT_CLEAN_WEIGHT=1.0, IMAGE_SIZE=[512,512],EFF_VER=7, BATCHSIZE=8*8=64
- Run 'v2total_train.ipynb' again, with larger image sizes.([640,640], [736,736])
- Run 'v2total_train.ipynb' again, with NOT_CLEAN_WEIGHT = 0.5
INFERENCE(./notebooks)
You can refer 'export_model.ipynb' for exporting single model and ensemble model.