Awesome
Quantifying Privacy Leakage in Graph Embedding (MobiQuitous'20 and NeurIPS PPML '20)
Preprint: https://arxiv.org/pdf/2010.00906.pdf
The code is as follows:
-
AIA: Attribute inference attacks. The embeddings are already generated using code form their original repos.
-
MIA: Membership Inference attacks. The code is divided into blackbox (includes confidence score attacks and shadow model attack) and whitebox. MIA in blackbox setting is performed on inductive training of GraphSage model.
-
Reconstruction: Graph Reconstruction using graph encoder decoder.
Data: The facebook and LastFM dataset for attribute inference attacks is available from Stanford Large Network Datasete Collection. The data for graph reconstruction requires to load the train and test graphs seperately unlike what most libraries provider. The data can be obtained from https://github.com/DaehanKim/vgae_pytorch.
Credits:
-
Node2Vec and DeepWalk embeddings algorithms from their original repositories (https://github.com/phanein/deepwalk) (https://github.com/aditya-grover/node2vec)
-
Graph Encoder Decoder: https://github.com/DaehanKim/vgae_pytorch
-
BlackBox MIA: https://github.com/inspire-group/privacy-vs-robustness