rickiepark/llm-from-scratch
<밑바닥부터 만들면서 공부하는 LLM>(길벗, 2025)의 코드 저장소
This project provides the official code repository for the book "LLM From Scratch" (Gilbut, 2025). It helps you understand how large language models (LLMs) like GPT work internally by guiding you through coding them step-by-step from the ground up. You'll input explanations, diagrams, and examples, and output a functional, albeit small, LLM that you can pre-train and fine-tune. This is for anyone, particularly data scientists, machine learning engineers, or researchers, who wants a deep, hands-on understanding of LLM architecture and development.
Use this if you want to build and understand large language models from their fundamental components without relying on high-level libraries.
Not ideal if you need to use existing, large-scale LLMs for practical applications or if you are not comfortable with Python programming.
Stars
97
Forks
108
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
Dec 16, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/rickiepark/llm-from-scratch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
AI-Hypercomputer/maxtext
A simple, performant and scalable Jax LLM!
rasbt/reasoning-from-scratch
Implement a reasoning LLM in PyTorch from scratch, step by step
mindspore-lab/mindnlp
MindSpore + 🤗Huggingface: Run any Transformers/Diffusers model on MindSpore with seamless...
mosaicml/llm-foundry
LLM training code for Databricks foundation models
CASE-Lab-UMD/LLM-Drop
The official implementation of the paper "Uncovering the Redundancy in Transformers via a...