SonySemiconductorSolutions/mct-model-optimization
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
Deploying neural networks on devices with limited computational power can be challenging. This tool helps optimize your pre-trained PyTorch or Keras models by reducing their size and computational demands, making them efficient for edge devices. It takes your existing floating-point model and outputs a compressed, quantized model suitable for deployment, benefiting AI/ML engineers and researchers working with resource-constrained hardware.
431 stars.
Use this if you need to deploy your neural network models on edge devices or hardware with limited memory and processing capabilities.
Not ideal if you are developing models for high-performance computing environments without strict hardware constraints, as the optimization process introduces complexity.
Stars
431
Forks
79
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 12, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/SonySemiconductorSolutions/mct-model-optimization"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related frameworks
open-mmlab/mmengine
OpenMMLab Foundational Library for Training Deep Learning Models
Xilinx/brevitas
Brevitas: neural network quantization in PyTorch
google/qkeras
QKeras: a quantization deep learning library for Tensorflow Keras
fastmachinelearning/qonnx
QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX
tensorflow/model-optimization
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization...