AttentionX/testGPT
Test-driven implementation of nanoGPT
This project provides a series of tested, step-by-step implementations for building large language models (LLMs) from scratch. It takes raw text data as input and produces a functional text generation model. This is for machine learning engineers, researchers, and students interested in understanding the fundamental building blocks of modern LLMs.
No commits in the last 6 months.
Use this if you are a machine learning practitioner looking to deeply understand and implement transformer-based language models from first principles.
Not ideal if you are looking for a pre-trained, ready-to-use LLM for immediate application or a high-level API for model integration.
Stars
16
Forks
3
Language
Python
License
—
Category
Last pushed
Dec 05, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/AttentionX/testGPT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
vixhal-baraiya/microgpt-c
The most atomic way to train and inference a GPT in pure, dependency-free C
milanm/AutoGrad-Engine
A complete GPT language model (training and inference) in ~600 lines of pure C#, zero dependencies
LeeSinLiang/microGPT
Implementation of GPT from scratch. Design to be lightweight and easy to modify.
dubzdubz/microgpt-ts
A complete GPT built from scratch in TypeScript with zero dependencies
biegehydra/NanoGptDotnet
A miniature large language model (LLM) that generates shakespeare like text written in C#....