Skip to content

Latest commit

 

History

History
30 lines (24 loc) · 1.35 KB

README.md

File metadata and controls

30 lines (24 loc) · 1.35 KB

Graph-JEPA

Paper (arxiv)

Graph-level Representation Learning with Joint-Embedding Predictive Architectures

Python environment setup with Conda

conda create --name graphjepa python=3.8
conda activate graphjepa
conda install pytorch==1.12.1 torchvision==0.13.1 torchaudio==0.12.1 cudatoolkit=11.3 -c pytorch
conda install -c pyg pytorch-sparse
conda install -c pyg pytorch-scatter
conda install -c pyg pytorch-cluster
conda install -c pyg pyg
pip install yacs
pip install tensorboard
pip install networkx
pip install einops
pip install metis

Running Graph-JEPA

The code for all the available datasets used in the paper is available under the train/ folder. Inside train/configs you can find the specific configuration files used for each dataset, corresponding with Table 8 of the paper. launch.sh contains two examples of launch scripts that you can use to directly modify the default config. You can see a detailed list of the arguments in core/config.py

Reproducibility

For the sake of openess, we provide the training logs containing the results published in the paper in paper_logs/

Credits

This repository is largely based on Graph-ViT-MLPMixer, check out their work if you are interested in graph representation learning with Transformers.