Home

Awesome

Paper Build-and-Test Code Quality Python versions license: MIT

Investigating Generalization Behaviours of Generative Flow Networks (GFlowNets, GFNs)

This repo is modified and based off the implementation: gflownet - it contains GFlowNet-related training and environment code on graphs for investigating the generalization capabilities of GFlowNets.

GFlowNets and generalization

It is hypothesize that GFlowNets leverage the generalization potential of deep neural networks to assign probability mass to unvisited areas of the state space. This repo contains a graph generation benchmark environment (considering several rewards of varying difficulty) such that $p(x)$ can be tractably computed over all graphs. This set of graph generation tasks are used to assess the ability of GFlowNets to generalize and approximate $p(x)$ for unvisited areas of the state space. See this http link for the paper.

GFlowNets

GFlowNet, short for Generative Flow Network (sometimes also abbreviated as GFN), is a novel generative modelling framework for learning unnormalized probability mass functions over discrete spaces, particularly suited for discrete/combinatorial objects. Here, the focus is on graph generation.

The idea behind GFlowNets is to estimate flows in a (graph-theoretic) directed acyclic network. The network represents all possible ways of constructing an object, and so knowing the flow gives us a policy which we can follow to sequentially construct objects. Such a sequence of partially constructed objects is a trajectory. Perhaps confusingly, the network in a GFlowNet refers to the state space, not a neural network architecture. Here the objects we construct are themselves graphs, which are constructed node by node. To make policy predictions, we use a graph neural network, parameterizing the forward policy $P_F(s' | s; \theta)$. This GNN outputs per-node logits (e.g. add a node to the current graph, or add an edge between these two nodes), as well as per-graph logits (e.g. stop/"done constructing this object").

Citing

If you find this code useful in your research, please cite the following paper (expand for BibTeX):

<details> <summary> L. Atanackovic, E. Bengio. Investigating Generalization Behaviours of Generative Flow Networks, 2024. </summary>
@article{atanackovic2024,
  title={Investigating Generalization Behaviours of Generative Flow Networks},
  author={Atanackovic, Lazar and Bengio, Emmanuel},
  journal={arXiv preprint arXiv:2402.05309},
  year={2024}
}
</details>

Repo overview

Structure of repo:

See implementation notes for more.

Getting started

First, generate a cache of all states/graphs up to 7 nodes. To do this run the following:

python basic_graph_task.py --recompute-all ./data/basic_graphs 7

To train a single model on the graph benchmark tasks, run:

python expts/task_single_run_gfn.py

To run an experiment, e.g. training the distilled flow models for $F$ and $P_F$ over 3 seeds, use task_distilled_flows.py.

Installation

PIP

This package is installable as a PIP package, but since it depends on some torch-geometric package wheels, the --find-links arguments must be specified as well:

pip install -e . --find-links https://data.pyg.org/whl/torch-1.13.1+cu117.html

Or for CPU use:

pip install -e . --find-links https://data.pyg.org/whl/torch-1.13.1+cpu.html

To install or depend on a specific tag, for example here v0.0.10, use the following scheme:

pip install git+https://github.com/recursionpharma/gflownet.git@v0.0.10 --find-links ...

If package dependencies seem not to work, you may need to install the exact frozen versions listed requirements/, i.e. pip install -r requirements/main_3.9.txt.