akshat0123/GPT-1
Pytorch implementation of GPT-1
This project helps machine learning engineers or researchers build and train a foundational language model like GPT-1 from scratch. It takes large text datasets, preprocesses them into a tokenized format, and outputs a trained model capable of understanding and generating human-like text. The primary users are those working on natural language processing tasks who need a customizable language model.
No commits in the last 6 months.
Use this if you are a machine learning engineer or researcher who needs to implement, pre-train, or fine-tune a GPT-1 architecture for specific natural language processing applications.
Not ideal if you are looking for an out-of-the-box solution to generate text or perform NLP tasks without needing to delve into model training and architecture.
Stars
35
Forks
7
Language
Python
License
—
Category
Last pushed
May 28, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/akshat0123/GPT-1"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
AliHaiderAhmad001/GPT-from-Scratch-with-Tensorflow
Implementation for "Improving Language Understanding by Generative Pre-Training" paper
HomebrewML/HomebrewNLP-torch
A case study of efficient training of large language models using commodity hardware.
qiqiApink/MotionGPT
The official PyTorch implementation of the paper "MotionGPT: Finetuned LLMs are General-Purpose...
nawnoes/pytorch-gpt-x
An implementation of an autoregressive language model using an improved Transformer and...
Shenggan/atp
Adaptive Tensor Parallelism for Foundation Models