Published March 25, 2023 | Version v2
Other Open

Betty: Enabling Large Scale GNN Training with Batch Level Graph Partitioning

  • 1. University of California, Merced
  • 2. Microsoft Research

Contributors

Producer:

  • 1. University of California, Merced

Description

This artifact includes the source codes and expected experimental data for replicating the evaluations in this paper.
We implement figure 2 to denote the OOM situation of current advanced GNN training, and applied figure 10 to illustrate Betty break the memory wall. We implement memory consumption estimation during the workflow of Betty, shown in figure 5. We use figure 12 to denote the tendency of peak memory consumption and training time per epoch as the number of micro batches increases. And the model convergence is not impacted by Betty and micro-batch training can be proved by the figure 13.
The framework of Betty is developed upon DGL(pytorch backend). The requirements: pytorch >= 1.7, DGL >= 0.7. The other software dependency include sortedcontainers, pyvis, pynvml, tqdm, pymetis, seaborn.
Our experiments result denoted in paper were collected from the machine with a RTX6000 GPU(24 GB memory) and Intel(R) Xeon(R) Gold 6126 CPU @ 2.60GHz. You can use a different configuration with at least one GPU.

Files

Betty-master (1).zip

Files (523.0 kB)

Name Size Download all
md5:942343ada28c514677fa8c3c7c4ca1f2
523.0 kB Preview Download

Additional details

References

  • Wang, M. Y. (2019, January). Deep graph library: Towards efficient and scalable deep learning on graphs. In ICLR workshop on representation learning on graphs and manifolds.