maxencefaldor/nanoGPT-JAX
Repository nanoGPT from @karpathy accelerated with JAX/Flax! The simplest, fastest repository for training/finetuning medium-sized GPTs.
This project helps machine learning engineers quickly train or fine-tune medium-sized GPT models. You provide a dataset of text, and it generates a functional GPT model capable of producing new, similar text outputs. It's designed for ML engineers who need to efficiently work with generative pre-trained transformers.
No commits in the last 6 months.
Use this if you are a machine learning engineer looking for a simple and fast way to train or fine-tune medium-sized GPT models using JAX/Flax.
Not ideal if you are looking for an out-of-the-box text generation tool without needing to engage in model training or fine-tuning.
Stars
9
Forks
2
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Jun 06, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/maxencefaldor/nanoGPT-JAX"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
vixhal-baraiya/microgpt-c
The most atomic way to train and inference a GPT in pure, dependency-free C
milanm/AutoGrad-Engine
A complete GPT language model (training and inference) in ~600 lines of pure C#, zero dependencies
LeeSinLiang/microGPT
Implementation of GPT from scratch. Design to be lightweight and easy to modify.
dubzdubz/microgpt-ts
A complete GPT built from scratch in TypeScript with zero dependencies
biegehydra/NanoGptDotnet
A miniature large language model (LLM) that generates shakespeare like text written in C#....