AlibabaPAI/FlashModels
Fast and easy distributed model training examples.
This tool helps AI engineers and researchers efficiently train large language models using distributed computing. You provide a pre-existing large language model configuration and training data, and it outputs a trained model, often much faster than standard methods. It's designed for those working with large-scale deep learning.
No commits in the last 6 months.
Use this if you need to train large language models like LLaMA-2 or QWen more quickly and efficiently across multiple GPUs or TPUs.
Not ideal if you are looking for a tool for general machine learning tasks or models outside of the large language model domain, or if you don't have access to distributed computing resources.
Stars
12
Forks
4
Language
Python
License
Apache-2.0
Category
Last pushed
Nov 26, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/AlibabaPAI/FlashModels"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
deepspeedai/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference...
helmholtz-analytics/heat
Distributed tensors and Machine Learning framework with GPU and MPI acceleration in Python
hpcaitech/ColossalAI
Making large AI models cheaper, faster and more accessible
horovod/horovod
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.
bsc-wdc/dislib
The Distributed Computing library for python implemented using PyCOMPSs programming model for HPC.