chester256/Model-Compression-Papers

Papers for deep neural network compression and acceleration

41
/ 100
Emerging

This collection helps machine learning engineers and researchers optimize deep learning models for better performance on resource-constrained devices. It provides a curated list of research papers on techniques like quantization, pruning, and knowledge distillation. The input is an existing neural network model and the output is guidance on how to make that model smaller, faster, and more efficient.

401 stars. No commits in the last 6 months.

Use this if you need to deploy large neural networks on edge devices, mobile phones, or embedded systems with limited computational power and memory.

Not ideal if you are looking for ready-to-use code implementations or a general overview of deep learning without a focus on model efficiency.

model-optimization deep-learning-deployment edge-ai neural-network-efficiency resource-constrained-ml
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 23 / 25

How are scores calculated?

Stars

401

Forks

80

Language

License

Last pushed

Jun 21, 2021

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/chester256/Model-Compression-Papers"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.