nawnoes/pytorch-gpt-x

An implementation of an autoregressive language model using an improved Transformer and DeepSpeed pipeline parallelism.

30
/ 100
Emerging

This project helps machine learning researchers or engineers train large language models (like GPT-style models) on limited hardware. You provide text data, and it trains a ~1 billion parameter model optimized with techniques like ReZero and DeepSpeed, enabling efficient training on just two V100 16GB GPUs. This is for individuals or teams developing advanced natural language processing capabilities.

Use this if you need to train a large autoregressive language model efficiently on a cluster with a small number of powerful GPUs.

Not ideal if you're looking for an off-the-shelf pre-trained model or if you don't have access to specialized GPU hardware.

large-language-models deep-learning-training natural-language-generation ml-research gpu-optimization
No License No Package No Dependents
Maintenance 6 / 25
Adoption 7 / 25
Maturity 8 / 25
Community 9 / 25

How are scores calculated?

Stars

30

Forks

3

Language

Python

License

Last pushed

Jan 12, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/nawnoes/pytorch-gpt-x"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.