draktr/findi-descent
FinDi: Finite Difference Gradient Descent can optimize any function, including the ones without analytic form, by employing finite difference numerical differentiation within a gradient descent algorithm
No commits in the last 6 months. Available on PyPI.
Stars
—
Forks
—
Language
Python
License
MIT
Category
Last pushed
Sep 14, 2024
Monthly downloads
13
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/draktr/findi-descent"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
jolars/sortedl1
Python package for Sorted L-One Penalized Estimation (SLOPE)
gugarosa/opytimizer
🐦 Opytimizer is a Python library consisting of meta-heuristic optimization algorithms.
SENATOROVAI/gradient-descent-sgd-solver-course
Stochastic Gradient Descent (SGD) is an optimization algorithm that updates model parameters...
hiroyuki-kasai/SGDLibrary
MATLAB/Octave library for stochastic optimization algorithms: Version 1.0.20
softmin/ReHLine-python
Regularized Composite ReLU-ReHU Loss Minimization with Linear Computation and Linear Convergence