HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism

Yanping Huang; Youlong Cheng; Ankur Bapna; Orhan Firat; Mia Xu Chen; Dehao Chen; HyoukJoong Lee; Jiquan Ngiam; Quoc V. Le; Yonghui Wu; Zhifeng Chen

GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism

Abstract

Scaling up deep neural network capacity has been known as an effective approach to improving model quality for several different machine learning tasks. In many cases, increasing model capacity beyond the memory limit of a single accelerator has required developing special algorithms or infrastructure. These solutions are often architecture-specific and do not transfer to other tasks. To address the need for efficient and task-independent model parallelism, we introduce GPipe, a pipeline parallelism library that allows scaling any network that can be expressed as a sequence of layers. By pipelining different sub-sequences of layers on separate accelerators, GPipe provides the flexibility of scaling a variety of different networks to gigantic sizes efficiently. Moreover, GPipe utilizes a novel batch-splitting pipelining algorithm, resulting in almost linear speedup when a model is partitioned across multiple accelerators. We demonstrate the advantages of GPipe by training large-scale neural networks on two different tasks with distinct network architectures: (i) Image Classification: We train a 557-million-parameter AmoebaNet model and attain a top-1 accuracy of 84.4% on ImageNet-2012, (ii) Multilingual Neural Machine Translation: We train a single 6-billion-parameter, 128-layer Transformer model on a corpus spanning over 100 languages and achieve better quality than all bilingual models.

Code Repositories

alondj/Pytorch-Gpipe
pytorch
Mentioned in GitHub
pikkaay/efficientnet_gpu
tf
Mentioned in GitHub
qubvel/efficientnet
tf
Mentioned in GitHub
xslidi/EfficientNets_ddl_apex
pytorch
Mentioned in GitHub
KakaoBrain/torchgpipe
pytorch
Mentioned in GitHub
northeastsquare/effficientnet
tf
Mentioned in GitHub
pytorch/pippy
pytorch
Mentioned in GitHub
shijianjian/efficientnet-pytorch-3d
pytorch
Mentioned in GitHub
yakhyo/EfficientNet-PyTorch
pytorch
Mentioned in GitHub
pytorch/tau
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
fine-grained-image-classification-on-birdsnapGPIPE
Accuracy: 83.6%
fine-grained-image-classification-on-stanfordGPipe
Accuracy: 94.6%
image-classification-on-cifar-10GPIPE + transfer learning
Percentage correct: 99
image-classification-on-cifar-100GPIPE
Percentage correct: 91.3
image-classification-on-imagenetGPIPE
Top 1 Accuracy: 84.4%

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp