mytechnotalent/kgpt

A GPT-2-class language model trained from scratch on OpenWebText with the intent to augment AI Transformer-model education and reverse engineer GPT models from scratch.

45
/ 100
Emerging

This project helps AI researchers and students learn how large language models work by allowing them to build one from scratch. You start with raw web text, process it, and train a GPT-2 class model. The result is a foundational model that can be further fine-tuned into a conversational chatbot.

Use this if you are an AI researcher or student eager to understand, reverse engineer, and experiment with the core mechanics of GPT-like transformer models from the ground up.

Not ideal if you're looking for an off-the-shelf, pre-trained language model for immediate application in a production environment.

AI Education Transformer Models Natural Language Processing Machine Learning Research Deep Learning
No Package No Dependents
Maintenance 10 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 13 / 25

How are scores calculated?

Stars

24

Forks

4

Language

Jupyter Notebook

License

Apache-2.0

Last pushed

Mar 04, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/mytechnotalent/kgpt"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.