hkproj/pytorch-transformer-distributed

Distributed training (multi-node) of a Transformer model

38
/ 100
Emerging

This project helps machine learning engineers train large transformer models more efficiently by distributing the computational load across multiple GPU-enabled machines. You provide your transformer model code and training data, and it outputs a trained model faster than single-machine setups. This is for machine learning engineers and researchers working with substantial AI models.

No commits in the last 6 months.

Use this if you need to accelerate the training of a large transformer model that is too computationally intensive for a single GPU machine.

Not ideal if you are training smaller models or do not have access to a multi-node, multi-GPU cloud computing environment.

deep-learning model-training natural-language-processing distributed-computing AI-research
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 8 / 25
Community 21 / 25

How are scores calculated?

Stars

94

Forks

40

Language

Python

License

Last pushed

Apr 10, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/hkproj/pytorch-transformer-distributed"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.