AdrianBZG/LLM-distributed-finetune

Tune efficiently any LLM model from HuggingFace using distributed training (multiple GPU) and DeepSpeed. Uses Ray AIR to orchestrate the training on multiple AWS GPU instances

35
/ 100
Emerging

This project helps machine learning engineers efficiently customize large language models (LLMs) like FALCON-7B for specific tasks or languages. You provide a pre-trained HuggingFace LLM and a dataset of examples, and it outputs a fine-tuned model ready for deployment. This is for ML engineers working with powerful GPU clusters on cloud platforms like AWS.

No commits in the last 6 months.

Use this if you need to quickly and efficiently fine-tune a HuggingFace Large Language Model using multiple GPUs and distributed training on AWS.

Not ideal if you do not have access to AWS GPU instances or prefer to fine-tune models on a single machine.

large-language-models model-fine-tuning distributed-training cloud-ml-ops natural-language-processing
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

60

Forks

6

Language

Python

License

MIT

Last pushed

Jun 20, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/AdrianBZG/LLM-distributed-finetune"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.