YuchenJin/autolrs
Automatic learning-rate scheduler
This project helps machine learning engineers or researchers automatically adjust the learning rate during deep neural network training. Instead of manually experimenting to find the best learning rate schedule, you provide your model and training data, and the system intelligently tunes the learning rate on the fly. This results in faster and more efficient training of deep learning models.
No commits in the last 6 months.
Use this if you are training deep neural networks and want to automate the complex process of finding an optimal learning rate schedule to improve training speed and model performance.
Not ideal if your training process involves a custom warmup phase where a specific learning rate behavior is intentionally desired and not based on minimizing validation loss.
Stars
46
Forks
11
Language
Python
License
MIT
Category
Last pushed
Apr 12, 2021
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/YuchenJin/autolrs"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
optuna/optuna
A hyperparameter optimization framework
keras-team/keras-tuner
A Hyperparameter Tuning Library for Keras
KernelTuner/kernel_tuner
Kernel Tuner
syne-tune/syne-tune
Large scale and asynchronous Hyperparameter and Architecture Optimization at your fingertips.
deephyper/deephyper
DeepHyper: A Python Package for Massively Parallel Hyperparameter Optimization in Machine Learning