Beomi/easy-lm-trainer
π€ μ΅μνμ μΈν μΌλ‘ LMμ νμ΅νκΈ° μν μνμ½λ
This project helps machine learning engineers and researchers quickly set up and train large language models using the Hugging Face Transformers library. You provide a dataset, and it outputs a trained language model ready for downstream tasks like text generation or summarization. It's designed for those who want to get a basic causal language model (CLM) training run operational with minimal initial configuration.
No commits in the last 6 months.
Use this if you are a machine learning engineer or researcher looking for a straightforward boilerplate to start training a causal language model with Hugging Face Transformers.
Not ideal if you need to fine-tune a model for a highly specialized task requiring complex custom training loops or advanced distributed training strategies beyond what Deepspeed offers out-of-the-box.
Stars
59
Forks
7
Language
Python
License
—
Category
Last pushed
May 23, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Beomi/easy-lm-trainer"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
unslothai/unsloth
Fine-tuning & Reinforcement Learning for LLMs. π¦₯ Train OpenAI gpt-oss, DeepSeek, Qwen, Llama,...
huggingface/peft
π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
modelscope/ms-swift
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs (Qwen3.5, DeepSeek-R1, GLM-5,...
oumi-ai/oumi
Easily fine-tune, evaluate and deploy gpt-oss, Qwen3, DeepSeek-R1, or any open source LLM / VLM!
linkedin/Liger-Kernel
Efficient Triton Kernels for LLM Training