GeeeekExplorer/transformers-patch
patches for huggingface transformers to save memory
This helps AI engineers and researchers reduce the memory footprint when working with large language models built using HuggingFace Transformers. By simply importing a patch, you can load and run larger models or process longer sequences of text with the same GPU resources, avoiding 'out of memory' errors. It takes existing Transformers models as input and allows them to operate more efficiently within your available hardware.
No commits in the last 6 months.
Use this if you are an AI engineer or researcher experiencing GPU memory limitations when running or fine-tuning large language models from HuggingFace Transformers.
Not ideal if you are not working with HuggingFace Transformers models or if your primary bottleneck is not GPU memory.
Stars
35
Forks
4
Language
Python
License
MIT
Category
Last pushed
Jun 02, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/GeeeekExplorer/transformers-patch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
AI-Hypercomputer/maxtext
A simple, performant and scalable Jax LLM!
rasbt/reasoning-from-scratch
Implement a reasoning LLM in PyTorch from scratch, step by step
mindspore-lab/mindnlp
MindSpore + 🤗Huggingface: Run any Transformers/Diffusers model on MindSpore with seamless...
mosaicml/llm-foundry
LLM training code for Databricks foundation models
rickiepark/llm-from-scratch
<밑바닥부터 만들면서 공부하는 LLM>(길벗, 2025)의 코드 저장소