mohamedrxo/simplegrad
SimpleGrad – a minimal NumPy-based autograd and neural network framework for learning automatic differentiation and basic deep learning concepts.
This library helps developers understand how automatic differentiation, a core concept in machine learning, works from the ground up. It takes numerical data as input and allows users to build simple neural networks and compute gradients for learning. It's designed for software developers, students, or researchers who are learning the foundational mechanics of deep learning frameworks.
No commits in the last 6 months.
Use this if you are a developer learning the underlying principles of machine learning libraries like PyTorch or TensorFlow.
Not ideal if you are looking for a high-performance, production-ready machine learning framework for building complex models.
Stars
8
Forks
—
Language
Jupyter Notebook
License
—
Category
Last pushed
Sep 28, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/mohamedrxo/simplegrad"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
JonathanRaiman/theano_lstm
:microscope: Nano size Theano LSTM module
google/tangent
Source-to-Source Debuggable Derivatives in Pure Python
ahrefs/ocannl
OCANNL: OCaml Compiles Algorithms for Neural Networks Learning
yoshoku/numo-openblas
Numo::OpenBLAS builds and uses OpenBLAS as a background library for Numo::Linalg
statusfailed/catgrad
a categorical deep learning compiler