Home

Awesome

PyTorch implementation of DistillGCN

Paper: Distilling Knowledge From Graph Convolutional Networks, CVPR'20

Method Overview

Dependencies

Main packages

PyTorch = 1.1.0

DGL = 1.4.0

See requirment file for more information about how to install the dependencies.

Training and evaluation

The main.py file contains the code for training teacher model, training the student model using the LSP module.

Early stop is used when training both the student model and the teacher model.

Cite

@inproceedings{yang2020distilling,
  title={Distilling Knowledge From Graph Convolutional Networks},
  author={Yang, Yiding and Qiu, Jiayan and Song, Mingli and Tao, Dacheng and Wang, Xinchao},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={7074--7083},
  year={2020}
}

License

DistillGCN is released under the MIT license. Please see the LICENSE file for more information.