kakaobrain/torchgpipe

A GPipe implementation in PyTorch

55
/ 100
Established

This tool helps machine learning engineers train extremely large neural networks that might otherwise exceed the memory capacity of a single GPU. It takes your existing PyTorch model and training data, then intelligently splits the model and data across multiple GPUs. The output is a successfully trained, massive model, enabling research and development of state-of-the-art AI.

862 stars. No commits in the last 6 months. Available on PyPI.

Use this if you are a machine learning engineer or researcher encountering 'out of memory' errors when trying to train very large PyTorch models on GPUs.

Not ideal if your models are small enough to train comfortably on a single GPU or if you are not using PyTorch and CUDA-enabled devices.

deep-learning neural-network-training large-scale-ml gpu-optimization
Stale 6m
Maintenance 0 / 25
Adoption 10 / 25
Maturity 25 / 25
Community 20 / 25

How are scores calculated?

Stars

862

Forks

98

Language

Python

License

BSD-3-Clause

Last pushed

Jul 25, 2024

Commits (30d)

0

Dependencies

1

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/kakaobrain/torchgpipe"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.