cgao96/nanoGPT.jax

The simplest, fastest repository for training/finetuning medium-sized GPTs in Jax.

29
/ 100
Experimental

This project helps machine learning engineers and researchers quickly train or finetune medium-sized GPT models on custom text data. You provide a dataset of text, and it outputs a trained language model capable of generating new text in a similar style or domain. It's designed for individuals who want to experiment with or develop their own text generation models.

No commits in the last 6 months.

Use this if you need a straightforward, fast, and hackable codebase to train a GPT-like model from scratch or adapt an existing one for specific text generation tasks.

Not ideal if you're looking for pre-built, production-ready GPT-2 models or an out-of-the-box solution for complex, large-scale natural language understanding tasks.

natural-language-generation text-synthesis custom-language-models machine-learning-research
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 4 / 25
Maturity 16 / 25
Community 9 / 25

How are scores calculated?

Stars

7

Forks

1

Language

Python

License

MIT

Last pushed

Jun 18, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/cgao96/nanoGPT.jax"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.