berkerdemirel/GPT-from-scratch

Re-implementation of Andrej Karpathy's nanoGPT

30
/ 100
Emerging

This project helps machine learning engineers and researchers understand how large language models (LLMs) work by providing a simplified, re-implemented version of a generative pre-trained transformer (GPT). You feed it text data, and it learns to generate new, similar text based on the patterns it identified. It's ideal for those looking to explore the foundational architecture of models like GPT.

No commits in the last 6 months.

Use this if you are a machine learning practitioner or student who wants to grasp the core mechanics of GPT models from the ground up, using a minimalist codebase.

Not ideal if you need a production-ready LLM for complex applications or a tool that solves a specific business problem beyond educational exploration.

AI-research NLP-development generative-AI-education deep-learning-engineering transformer-architecture
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 8 / 25
Community 16 / 25

How are scores calculated?

Stars

16

Forks

6

Language

Python

License

Last pushed

Feb 16, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/berkerdemirel/GPT-from-scratch"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.