GRAAL-Research/poutyne
A simplified framework and utilities for PyTorch
This framework helps machine learning practitioners efficiently train neural networks built with PyTorch. It takes your PyTorch network and data, then outputs a trained model, performance metrics, and predictions without needing to write extensive boilerplate code. Data scientists and machine learning engineers who use PyTorch for deep learning model development will find this useful.
579 stars. Used by 2 other packages. No commits in the last 6 months. Available on PyPI.
Use this if you are a machine learning engineer or data scientist working with PyTorch and want to streamline the process of training, evaluating, and managing your deep learning models.
Not ideal if you prefer to write all training loops and model management logic from scratch or are not working with PyTorch.
Stars
579
Forks
64
Language
Python
License
LGPL-3.0
Category
Last pushed
May 05, 2025
Commits (30d)
0
Dependencies
3
Reverse dependents
2
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/GRAAL-Research/poutyne"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related frameworks
pytorch/pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
keras-team/keras
Deep Learning for humans
Lightning-AI/torchmetrics
Machine learning metrics for distributed, scalable PyTorch applications.
Lightning-AI/pytorch-lightning
Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.
lanpa/tensorboardX
tensorboard for pytorch (and chainer, mxnet, numpy, ...)