saqib1707/gpt2-from-scratch
PyTorch Implementation of GPT-2
This project helps machine learning researchers and students understand how large language models work by providing a clear, simplified version of the GPT-2 model. You can take a high-quality educational text dataset, train the model, and then generate new text based on prompts you provide. This is designed for those learning about neural networks and natural language processing.
No commits in the last 6 months.
Use this if you are a student or researcher wanting to learn the inner workings of a GPT-2 model by running and experimenting with it yourself.
Not ideal if you need a production-ready, highly optimized text generation tool for real-world applications or fast inference.
Stars
31
Forks
11
Language
Python
License
MIT
Category
Last pushed
Sep 04, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/saqib1707/gpt2-from-scratch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
tabularis-ai/be_great
A novel approach for synthesizing tabular data using pretrained large language models
EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron...
shibing624/textgen
TextGen: Implementation of Text Generation models, include LLaMA, BLOOM, GPT2, BART, T5, SongNet...
ai-forever/ru-gpts
Russian GPT3 models.
AdityaNG/kan-gpt
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold...