BaguaSys/bagua

Bagua Speeds up PyTorch

53
/ 100
Established

This project helps machine learning engineers and researchers accelerate the training of their deep learning models on PyTorch. It takes your existing PyTorch model and training script, and through various optimizations like distributed training across multiple GPUs or machines, faster data loading, and communication enhancements, produces a significantly faster training process. The primary users are those working with large-scale deep learning models, especially in fields like computer vision or natural language processing, who need to reduce training times.

884 stars. No commits in the last 6 months. Available on PyPI.

Use this if you are a machine learning engineer or researcher using PyTorch and need to drastically speed up the training time of your deep learning models, particularly when working with large datasets or models that require multi-GPU or multi-machine setups.

Not ideal if you are not using PyTorch, or if your deep learning models are small and train quickly on a single GPU, as the overhead may outweigh the benefits.

deep-learning-training pytorch-optimization distributed-machine-learning model-training-acceleration large-scale-ai
Stale 6m No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 25 / 25
Community 18 / 25

How are scores calculated?

Stars

884

Forks

81

Language

Python

License

MIT

Last pushed

Aug 01, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/BaguaSys/bagua"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.