AlibabaPAI/FlashModels

Fast and easy distributed model training examples.

36
/ 100
Emerging

This tool helps AI engineers and researchers efficiently train large language models using distributed computing. You provide a pre-existing large language model configuration and training data, and it outputs a trained model, often much faster than standard methods. It's designed for those working with large-scale deep learning.

No commits in the last 6 months.

Use this if you need to train large language models like LLaMA-2 or QWen more quickly and efficiently across multiple GPUs or TPUs.

Not ideal if you are looking for a tool for general machine learning tasks or models outside of the large language model domain, or if you don't have access to distributed computing resources.

large-language-models distributed-training deep-learning-engineering AI-research model-optimization
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 16 / 25
Community 15 / 25

How are scores calculated?

Stars

12

Forks

4

Language

Python

License

Apache-2.0

Last pushed

Nov 26, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/AlibabaPAI/FlashModels"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.