BriefGPT.xyz
Aug, 2023
GNNPipe: 使用流水线模型并行加速分布式全图GNN训练
GNNPipe: Accelerating Distributed Full-Graph GNN Training with Pipelined Model Parallelism
HTML
PDF
Jingji Chen, Zhuoming Chen, Xuehai Qian
TL;DR
采用模型并行而非图并行的分布式全局图神经网络训练方法GNNPipe,结合基于分块的流水线训练方法以及混合并行性,以减少通信开销并加快训练时间,同时保持相当的模型准确性和收敛速度。
Abstract
Current
distributed full-graph gnn training
methods adopt a variant of data parallelism, namely
graph parallelism
, in which the whole graph is divided into multiple partitions (subgraphs) and each GPU processes o
→