Shenggan/atp

Adaptive Tensor Parallelism for Foundation Models

21
/ 100
Experimental

This project helps machine learning engineers and researchers efficiently train and deploy very large AI models, often called foundation models. It takes your existing large model architecture and training setup, then intelligently optimizes how computations are distributed across multiple GPUs or machines. The result is faster training times and more efficient inference for your large AI models.

No commits in the last 6 months.

Use this if you are working with extremely large AI models and need to reduce their training or inference time by optimizing how they utilize distributed hardware.

Not ideal if you are working with smaller models or do not have access to a distributed computing environment with multiple GPUs.

large-language-models distributed-training model-deployment deep-learning-infrastructure AI-model-optimization
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 16 / 25
Community 0 / 25

How are scores calculated?

Stars

9

Forks

Language

Python

License

MIT

Last pushed

Dec 15, 2022

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Shenggan/atp"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.