aitechnologies-it/gpt-mini
Yet another minimalistic Tensorflow (re-)re-implementation of Karpathy's Pytorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer).
This project provides a foundational building block for AI research, allowing you to train a simplified Generative Pretrained Transformer (GPT) model using the TensorFlow framework. You can input text or image data and train the model to generate new, similar text or images. This is ideal for AI researchers, students, or educators exploring the mechanics of transformer models for generative tasks.
No commits in the last 6 months.
Use this if you are an AI researcher or student who wants to understand and experiment with the core components of a GPT model for text or image generation in a minimalistic TensorFlow environment.
Not ideal if you need a production-ready large language model or a tool for advanced, large-scale natural language processing applications.
Stars
14
Forks
—
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Nov 18, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/aitechnologies-it/gpt-mini"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
LowinLi/transformers-stream-generator
This is a text generation method which returns a generator, streaming out each token in...
ystemsrx/mini-nanoGPT
One-click training of your own GPT. Training a GPT has never been easier for beginners. /...
jaymody/picoGPT
An unnecessarily tiny implementation of GPT-2 in NumPy.
kyegomez/AttentionGrid
A network of attention mechanisms at your fingertips. Unleash the potential of attention...
kamalkraj/minGPT-TF
A minimal TF2 re-implementation of the OpenAI GPT training