abhaskumarsinha/MinimalGPT
MinimalGPT is a concise, adaptable, and streamlined code framework that encompasses the essential components necessary for the construction, training, inference, and fine-tuning of the GPT model. This framework is implemented exclusively using Keras and TensorFlow, ensuring compatibility and coherence within the broader deep learning ecosystem.
This is a framework for machine learning engineers and researchers to experiment with a foundational GPT model. You provide your own text corpus as input and can build, train, and fine-tune a GPT model to generate new text. This project is for those who need a stripped-down, essential toolkit to understand or modify the core components of GPT.
No commits in the last 6 months.
Use this if you are a deep learning practitioner who wants to build, train, or fine-tune a simple GPT model from scratch using Keras and TensorFlow.
Not ideal if you are looking for a pre-trained, production-ready language model or a high-level API for text generation.
Stars
24
Forks
6
Language
Python
License
MIT
Category
Last pushed
Apr 03, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/abhaskumarsinha/MinimalGPT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
LowinLi/transformers-stream-generator
This is a text generation method which returns a generator, streaming out each token in...
ystemsrx/mini-nanoGPT
One-click training of your own GPT. Training a GPT has never been easier for beginners. /...
jaymody/picoGPT
An unnecessarily tiny implementation of GPT-2 in NumPy.
kyegomez/AttentionGrid
A network of attention mechanisms at your fingertips. Unleash the potential of attention...
kamalkraj/minGPT-TF
A minimal TF2 re-implementation of the OpenAI GPT training