Tencent/PatrickStar
PatrickStar enables Larger, Faster, Greener Pretrained Models for NLP and democratizes AI for everyone.
This project helps AI engineers and researchers train very large language models without needing enormous computing resources. It takes your existing PyTorch model code and a configuration, allowing you to train models significantly larger than typically possible on your available GPUs by efficiently using both GPU and CPU memory. This is for machine learning practitioners working with large-scale natural language processing.
779 stars.
Use this if you are encountering 'out-of-memory' errors when trying to train large language models and want to use fewer GPUs or train much bigger models with your current hardware.
Not ideal if you are working with small models or have abundant GPU resources and do not face memory constraints.
Stars
779
Forks
59
Language
Python
License
BSD-3-Clause
Category
Last pushed
Nov 18, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Tencent/PatrickStar"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
OpenMotionLab/MotionGPT
[NeurIPS 2023] MotionGPT: Human Motion as a Foreign Language, a unified motion-language...
wenlu-lab/cMolGPT
GPT (Generative Pre-trained Transformer) for de novo molecular design by enforcing specified targets
OpenMotionLab/MotionGPT3
MotionGPT3: Human Motion as a Second Modality, a MoT-based framework for unified motion...
SmerkyG/gptcore
Fast modular code to create and train cutting edge LLMs
YerbaPage/DetectCodeGPT
Detection of LLM-Generated Codes [ICSE2025]