Awesome
Extended SimGNN
A PyTorch Geometric implementation of "SimGNN: A Neural Network Approach to Fast Graph Similarity Computation" (WSDM 2019) [Paper] extended with Graph Isomorphism Operator from the “How Powerful are Graph Neural Networks?” paper and Differentiable Pooling Operator from the "Hierarchical Graph Representation Learning with Differentiable Pooling" paper.
This implementation was written and used to conduct experiments for my bachelor thesis "Approximating the Graph Edit Distance via Deep Learning" at the Technical University of Dortmund.
The code in this repository can be changed and updated. The original code that was attached to my thesis can be viewed in the initial commit.
The project structure is based on the PyTorch implementation [benedekrozemberczki/SimGNN].
A reference Tensorflow implementation is accessible [here].
Requirements
The codebase is implemented in Python 3.6.9. package versions used for development are just below.
matplotlib 3.1.3
networkx 2.4
numpy 1.19.5
pandas 1.1.2
scikit-learn 0.23.2
scipy 1.4.1
texttable 1.6.3
torch 1.7.1
torch-cluster 1.5.9
torch-geometric 1.7.0
torch-scatter 2.0.6
torch-sparse 0.6.9
torchvision 0.8.2
tqdm 4.60.0
Datasets
The datasets are taken with the help of GEDDataset, where the databases specified in the original repository with GED-values are loaded. Currently AIDS700nef, LINUX, ALKANE and IMDBMulti databases are supported.
Options
Training a SimGNN model is handled by the src/main.py
script which provides the following command line arguments.
Input and output options
--dataset STR Name of the dataset to be used. Default is `AIDS700nef`.
--plot BOOL Plot mse values during the learning. Default is False.
Model options
--diffpool BOOL Differentiable pooling. Default is False.
--gnn-operator STR Type of gnn operator. Default is gcn.
--filters-1 INT Number of filter in 1st GNN layer. Default is 64.
--filters-2 INT Number of filter in 2nd GNN layer. Default is 32.
--filters-3 INT Number of filter in 3rd GNN layer. Default is 16.
--tensor-neurons INT Neurons in tensor network layer. Default is 16.
--bottle-neck-neurons INT Bottle neck layer neurons. Default is 16.
--bins INT Number of histogram bins. Default is 16.
--batch-size INT Number of pairs processed per batch. Default is 128.
--epochs INT Number of SimGNN training epochs. Default is 350.
--dropout FLOAT Dropout rate. Default is 0.
--learning-rate FLOAT Learning rate. Default is 0.001.
--weight-decay FLOAT Weight decay. Default is 10^-5.
--histogram BOOL Include histogram features. Default is False.
Other options
--save STR Store the model Default is None.
--load STR Load a pretrained model Default is None.
--synth BOOL Generate and add synth data to the samples. Default is False.
--measure-time BOOL Measure average calculation time for one graph pair. Default is False.
Examples
The following commands learn a neural network and score on the test set. Training a SimGNN model on the default dataset.
python src/main.py
<p align="center">
<img style="float: center;" src="simgnn_run.png">
</p>
Training a SimGNN model for a 1000 epochs with a batch size of 512.
python src/main.py --epochs 1000 --batch-size 512
Training a SimGNN on the LINUX dataset with histogram features, GIN instead of GCN and DiffPool instead of attention mechanism from the paper.
python src/main.py --dataset LINUX --histogram --gnn-operator gin --diffpool
Drawing up a graphic of mse values during the learning.
python src/main.py --plot
Increasing the learning rate and the dropout.
python src/main.py --learning-rate 0.01 --dropout 0.9