google/qkeras
QKeras: a quantization deep learning library for Tensorflow Keras
This tool helps machine learning engineers and researchers optimize deep learning models for deployment on resource-constrained hardware like edge devices. By applying quantization techniques to Keras models, it significantly reduces the memory footprint and computational cost of neural networks. You provide an existing Keras deep learning model, and it outputs a quantized version that runs more efficiently.
578 stars. Available on PyPI.
Use this if you need to deploy deep learning models on hardware with limited memory and processing power, such as embedded systems or specialized accelerators, without sacrificing too much accuracy.
Not ideal if your primary goal is rapid prototyping or if you are running models on powerful cloud GPUs where performance and memory constraints are not a critical concern.
Stars
578
Forks
109
Language
Python
License
Apache-2.0
Category
Last pushed
Feb 23, 2026
Commits (30d)
0
Dependencies
9
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/google/qkeras"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related frameworks
open-mmlab/mmengine
OpenMMLab Foundational Library for Training Deep Learning Models
Xilinx/brevitas
Brevitas: neural network quantization in PyTorch
fastmachinelearning/qonnx
QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX
tensorflow/model-optimization
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization...
lucidrains/vector-quantize-pytorch
Vector (and Scalar) Quantization, in Pytorch