cgao96/nanoGPT.jax
The simplest, fastest repository for training/finetuning medium-sized GPTs in Jax.
This project helps machine learning engineers and researchers quickly train or finetune medium-sized GPT models on custom text data. You provide a dataset of text, and it outputs a trained language model capable of generating new text in a similar style or domain. It's designed for individuals who want to experiment with or develop their own text generation models.
No commits in the last 6 months.
Use this if you need a straightforward, fast, and hackable codebase to train a GPT-like model from scratch or adapt an existing one for specific text generation tasks.
Not ideal if you're looking for pre-built, production-ready GPT-2 models or an out-of-the-box solution for complex, large-scale natural language understanding tasks.
Stars
7
Forks
1
Language
Python
License
MIT
Category
Last pushed
Jun 18, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/cgao96/nanoGPT.jax"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
vixhal-baraiya/microgpt-c
The most atomic way to train and inference a GPT in pure, dependency-free C
milanm/AutoGrad-Engine
A complete GPT language model (training and inference) in ~600 lines of pure C#, zero dependencies
LeeSinLiang/microGPT
Implementation of GPT from scratch. Design to be lightweight and easy to modify.
dubzdubz/microgpt-ts
A complete GPT built from scratch in TypeScript with zero dependencies
biegehydra/NanoGptDotnet
A miniature large language model (LLM) that generates shakespeare like text written in C#....