Back to Pytorch Geometric

Examples for Distributed Training

examples/multi_gpu/README.md

2.7.03.8 KB
Original Source

Examples for Distributed Training

Examples with NVIDIA GPUs

Examples with cuGraph

For the best performance with NVIDIA GPUs, we recommend using cuGraph. Refer to our installation guide for setup instructions and to the cuGraph-PyG examples for ready-to-run training scripts covering single-node, multi-node, and link-prediction workloads.

Examples with Pure PyTorch

ExampleScalabilityDescription
distributed_batching.pysingle-nodeGraph-level prediction on many small graphs (ogbg-molhiv) using DataLoader with DistributedSampler.
distributed_sampling.pysingle-nodeNode-level classification on a single large graph (Reddit) using NeighborLoader for multi-hop subgraph sampling.
distributed_sampling_multinode.pymulti-nodeTraining GNNs on a homogeneous graph with neighbor sampling on multiple nodes.
distributed_sampling_multinode.sbatchmulti-nodeSubmitting a training job to a Slurm cluster using distributed_sampling_multinode.py.
papers100m_gcn.pysingle-nodeTraining GNNs on the ogbn-papers100M homogeneous graph w/ ~1.6B edges.
papers100m_gcn_multinode.pymulti-nodeTraining GNNs on a homogeneous graph on multiple nodes.
pcqm4m_ogb.pysingle-nodeTraining GNNs for a graph-level regression task.
mag240m_graphsage.pysingle-nodeTraining GNNs on a large heterogeneous graph.
taobao.pysingle-nodeTraining link prediction GNNs on a heterogeneous graph.
model_parallel.pysingle-nodeModel parallelism by manually placing layers on each GPU.

Examples with Intel GPUs (XPUs)

ExampleScalabilityDescription
distributed_sampling_xpu.pysingle-node, multi-gpuTraining GNNs on a homogeneous graph with neighbor sampling.