AdrianBZG/LLM-distributed-finetune
Tune efficiently any LLM model from HuggingFace using distributed training (multiple GPU) and DeepSpeed. Uses Ray AIR to orchestrate the training on multiple AWS GPU instances
This project helps machine learning engineers efficiently customize large language models (LLMs) like FALCON-7B for specific tasks or languages. You provide a pre-trained HuggingFace LLM and a dataset of examples, and it outputs a fine-tuned model ready for deployment. This is for ML engineers working with powerful GPU clusters on cloud platforms like AWS.
No commits in the last 6 months.
Use this if you need to quickly and efficiently fine-tune a HuggingFace Large Language Model using multiple GPUs and distributed training on AWS.
Not ideal if you do not have access to AWS GPU instances or prefer to fine-tune models on a single machine.
Stars
60
Forks
6
Language
Python
License
MIT
Category
Last pushed
Jun 20, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/AdrianBZG/LLM-distributed-finetune"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
TsinghuaC3I/MARTI
A Framework for LLM-based Multi-Agent Reinforced Training and Inference
zjunlp/KnowLM
An Open-sourced Knowledgable Large Language Model Framework.
cli99/llm-analysis
Latency and Memory Analysis of Transformer Models for Training and Inference
tanyuqian/redco
NAACL '24 (Best Demo Paper RunnerUp) / MlSys @ NeurIPS '23 - RedCoast: A Lightweight Tool to...
stanleylsx/llms_tool
一个基于HuggingFace开发的大语言模型训练、测试工具。支持各模型的webui、终端预测,低参数量及全参数模型训练(预训练、SFT、RM、PPO、DPO)和融合、量化。