codiceSpaghetti/numpyGPT

A from-scratch GPT built with NumPy and Python’s standard library. No autograd, no frameworks: every layer is re-implemented with its own forward and backward pass. Gradients are computed manually, updates are transparent, and every operation is spelled out.

24
/ 100
Experimental

This project helps machine learning engineers and researchers understand how large language models work from the ground up. By providing a re-implementation of a GPT model using only NumPy, it explicitly shows every mathematical operation involved in neural networks, tokenization, and optimization. You input raw text data, and it outputs a trained language model capable of generating new text, while also visualizing the training process.

Use this if you are a machine learning student, researcher, or engineer who wants a deep, transparent understanding of transformer architecture and backpropagation without the abstractions of higher-level frameworks.

Not ideal if you need to train large-scale, production-ready language models efficiently or are looking for a high-performance deep learning library.

natural-language-processing deep-learning-education neural-network-architecture text-generation backpropagation
No License No Package No Dependents
Maintenance 6 / 25
Adoption 6 / 25
Maturity 7 / 25
Community 5 / 25

How are scores calculated?

Stars

20

Forks

1

Language

Python

License

Last pushed

Nov 23, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/codiceSpaghetti/numpyGPT"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.