suvojit-0x55aa/mixed-precision-pytorch
Training with FP16 weights in PyTorch
This project helps machine learning engineers train deep learning models faster and more efficiently using NVIDIA GPUs. By training models with half-precision (FP16) weights, it reduces the memory footprint of models and significantly decreases training time. This is beneficial for anyone building and iterating on large-scale deep learning models who needs to optimize their training workflow.
No commits in the last 6 months.
Use this if you are a machine learning engineer or researcher looking to speed up the training of your PyTorch models on NVIDIA GPUs while reducing memory consumption.
Not ideal if your deep learning models are small, you are not using an NVIDIA GPU, or you require absolute maximal numerical precision at all times.
Stars
81
Forks
19
Language
Python
License
WTFPL
Category
Last pushed
Aug 07, 2019
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/suvojit-0x55aa/mixed-precision-pytorch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
triton-inference-server/server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
gpu-mode/Triton-Puzzles
Puzzles for learning Triton
hailo-ai/hailo_model_zoo
The Hailo Model Zoo includes pre-trained models and a full building and evaluation environment
open-mmlab/mmdeploy
OpenMMLab Model Deployment Framework
hyperai/tvm-cn
TVM Documentation in Chinese Simplified / TVM 中文文档