dheeren-tejani/mini-lm-124m
Experimental GPT-2 scale (~124M param) LLM trained from scratch on Google Colab. Trained on C4, Cosmopedia/Alpaca/Python mix. Includes full training pipeline, mixed dataset loader with Colab-resilient checkpointing, and log analysis tools. Honest write-up of what went wrong.
Stars
—
Forks
—
Language
TypeScript
License
—
Category
Last pushed
Feb 19, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/dheeren-tejani/mini-lm-124m"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
affjljoo3581/GPT2
PyTorch Implementation of OpenAI GPT-2
gyunggyung/LFM2-KoEn-Tuning
Fine-tuning LFM2-1.2B for Korean-English bidirectional translation. GRPO+COMET & SFT Training,...
akanyaani/Illustrated_GPT2_With_Code
Explained GPT-2 Transformer model step by step with code.
b14ucky/Taco-LLMingway
Custom GPT Transformer architecture built from scratch in PyTorch. Trained on Taco Hemingway's...