kyegomez/TeraGPT

Train a production grade GPT in less than 400 lines of code. Better than Karpathy's verison and GIGAGPT

46
/ 100
Emerging

This project helps machine learning engineers and researchers quickly build and train very large language models, like those with hundreds of billions of parameters, from scratch. You provide your training data, and it outputs a ready-to-use GPT-style model. It's designed for those who want to experiment with large-scale language model architectures without the complexity of traditional enterprise frameworks.

Available on PyPI.

Use this if you need to train or fine-tune a large language model (GPT-3 size or larger) and prefer a simple, compact codebase for rapid development and scaling.

Not ideal if you are looking for a pre-trained model or a no-code solution for applying existing large language models.

Large Language Models Deep Learning NLP Research Model Training AI Prototyping
Maintenance 10 / 25
Adoption 6 / 25
Maturity 25 / 25
Community 5 / 25

How are scores calculated?

Stars

16

Forks

1

Language

Python

License

MIT

Last pushed

Feb 06, 2026

Commits (30d)

0

Dependencies

4

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/kyegomez/TeraGPT"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.