mosaicml/llm-foundry
LLM training code for Databricks foundation models
This project helps machine learning engineers and researchers efficiently build, customize, and deploy large language models. You can input raw text data or pre-existing models, and it outputs finetuned, evaluated, and ready-to-use LLMs for various applications. It's designed for those who need to quickly experiment with and adapt powerful language models to specific tasks or datasets.
4,397 stars. Available on PyPI.
Use this if you are a machine learning practitioner looking to train, finetune, evaluate, or deploy large language models (LLMs) with high efficiency and flexibility.
Not ideal if you are looking for an off-the-shelf application or a simple API to use pre-trained models without needing customization or deep technical control.
Stars
4,397
Forks
584
Language
Python
License
Apache-2.0
Category
Last pushed
Oct 27, 2025
Commits (30d)
0
Dependencies
22
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/mosaicml/llm-foundry"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
AI-Hypercomputer/maxtext
A simple, performant and scalable Jax LLM!
rasbt/reasoning-from-scratch
Implement a reasoning LLM in PyTorch from scratch, step by step
mindspore-lab/mindnlp
MindSpore + 🤗Huggingface: Run any Transformers/Diffusers model on MindSpore with seamless...
rickiepark/llm-from-scratch
<밑바닥부터 만들면서 공부하는 LLM>(길벗, 2025)의 코드 저장소
CASE-Lab-UMD/LLM-Drop
The official implementation of the paper "Uncovering the Redundancy in Transformers via a...