approx-ml/approx

Automatic quantization library

27
/ 100
Experimental

This tool helps machine learning engineers and researchers optimize their trained deep learning models. It takes an existing, trained neural network model and automatically reduces its size and computational requirements. The output is a more efficient, 'quantized' version of the original model that can run faster and with less memory, ideal for deployment on resource-constrained devices or for improving inference speed.

No commits in the last 6 months.

Use this if you need to deploy a trained machine learning model on hardware with limited memory or processing power, or if you want to speed up model inference without a significant loss in accuracy.

Not ideal if your primary concern is developing the initial model architecture or training it from scratch, as this tool focuses on post-training optimization.

deep-learning-deployment model-optimization edge-ai neural-network-inference model-compression
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 16 / 25
Community 6 / 25

How are scores calculated?

Stars

12

Forks

1

Language

Python

License

Apache-2.0

Last pushed

Aug 11, 2022

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/approx-ml/approx"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.