Skip to main content
Ctrl+K

GiGL

  • User Guide
  • API Reference
  • GitHub
  • User Guide
  • API Reference
  • GitHub

Section Navigation

Overview

  • Overview
  • GiGL Architecture
    • Config Populator
    • Data Preprocessor
    • Subgraph Sampler
    • Split Generator
    • Trainer
    • Inference

Getting Started

  • Quick Start
  • Cloud Setup Guide
  • Installation
  • Orchestration
  • Examples
    • Toy Example - Tabularized GiGL
    • Examples for Training and Inference on Link Prediction GNN models.
      • Cora Distributed Training Example
      • DBLP Distributed Training Example
    • Running the MAG240M experiments on your own GCP project
      • (Optional) Fetch MAG240M Data into your own project
      • MAG240M E2E Example

Config Guides

  • Resource Config Guide
  • Task Config Guide
  • Data Preprocessor Spec Guide

Public Resources

  • Docker Images
  • Public Assets

Trouble Shooting

  • FAQ.md
  • User Guide
  • Examples
  • Examples for Training and Inference on Link Prediction GNN models.

Examples for Training and Inference on Link Prediction GNN models.#

Homogeneous (CORA)#

We use the CORA dataset as an example for sampling against a homogeneous dataset.

homogeneous_inference.py and homogeneous_training.py are example inference and training loops for the CORA dataset, the MNIST of graph models, and available via the PyG Planetoid dataset.

You can follow along with cora.ipynb to run an e2e GiGL pipeline on the CORA dataset. It will guide you through running each component: config_populator -> data_preprocessor -> trainer -> inferencer

Heterogeneous (DBLP)#

We use use the DBLP dataset as an example for sampling against a heterogeneous dataset.

heterogeneous_inference.py and heterogeneous_training.py are example inference and training loops for the DBLP dataset. The DBLP dataset is avaialble at the PyG DBLP dataset.

You can follow along with dblp.ipynb to run an e2e GiGL pipeline on the DBLP dataset. It will guide you through running each component: config_populator -> data_preprocessor -> trainer -> inferencer

previous

Toy Example - Tabularized GiGL

next

Cora Distributed Training Example

On this page
  • Homogeneous (CORA)
  • Heterogeneous (DBLP)
Edit on GitHub

© Copyright 2025, Snap Inc.

Created using Sphinx 7.4.7.

Built with the PyData Sphinx Theme 0.16.1.