ModelTC/LightCompress

[EMNLP 2024 & AAAI 2026] A powerful toolkit for compressing large models including LLMs, VLMs, and video generative models.

64
/ 100
Established

This toolkit helps organizations make their large AI models, like those for generating text, images, or video, run more efficiently and use less memory. It takes your existing large AI model and outputs a smaller, faster version that maintains its original performance. This is for AI developers and MLOps engineers who need to deploy these large models more cost-effectively on various hardware.

688 stars. Actively maintained with 36 commits in the last 30 days.

Use this if you need to deploy large AI models (LLMs, VLMs, video generative models) and want to reduce their size and inference costs without significant performance loss.

Not ideal if you are a general user without experience in model deployment or if you need to compress very small, specialized models.

AI model deployment MLOps large language models computer vision models generative AI
No Package No Dependents
Maintenance 20 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 18 / 25

How are scores calculated?

Stars

688

Forks

72

Language

Python

License

Apache-2.0

Last pushed

Mar 11, 2026

Commits (30d)

36

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/ModelTC/LightCompress"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.