mcognetta/LotteryTickets.jl

Sparsify Your Flux Models

21
/ 100
Experimental

This project helps machine learning engineers and researchers optimize their neural networks by finding smaller, more efficient 'sparse subnetworks' without losing performance. It takes an existing Flux deep learning model and iteratively prunes away less important connections, producing a compact version that often runs faster and uses less memory. This is ideal for anyone working with deep neural networks who needs to deploy models more efficiently or explore the fundamental structure of neural networks.

No commits in the last 6 months.

Use this if you are building deep learning models in Flux.jl and want to make them smaller, faster, or more memory-efficient without sacrificing accuracy.

Not ideal if your models are already very small or if you are not using the Flux.jl deep learning framework.

deep-learning neural-network-optimization model-compression machine-learning-research
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 16 / 25
Community 0 / 25

How are scores calculated?

Stars

14

Forks

Language

Julia

License

MIT

Last pushed

Sep 20, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/mcognetta/LotteryTickets.jl"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.