EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
This is a specialized toolkit for researchers and engineers who need to train very large language models from scratch, or fine-tune existing ones, using substantial computational resources. It takes raw text data and configuration settings as input, and outputs a custom-trained language model capable of generating human-like text. This is for users operating at the cutting edge of AI, often in academic, industry, or government labs.
7,399 stars.
Use this if you are a researcher or engineer looking to train large-scale language models with billions of parameters using distributed training across multiple GPUs or high-performance computing clusters.
Not ideal if you are looking to run generic inference with existing models or train smaller models; in those cases, simpler libraries like Hugging Face's `transformers` are more appropriate.
Stars
7,399
Forks
1,100
Language
Python
License
Apache-2.0
Category
Last pushed
Feb 03, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/EleutherAI/gpt-neox"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related models
tabularis-ai/be_great
A novel approach for synthesizing tabular data using pretrained large language models
shibing624/textgen
TextGen: Implementation of Text Generation models, include LLaMA, BLOOM, GPT2, BART, T5, SongNet...
ai-forever/ru-gpts
Russian GPT3 models.
AdityaNG/kan-gpt
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold...
zemlyansky/gpt-tfjs
GPT in TensorFlow.js