windreamer/flash-attention3-wheels

Pre-built wheels that erase Flash Attention 3 installation headaches.

36
/ 100
Emerging

This project simplifies the installation of Flash Attention 3, a highly optimized component used in large language models and other deep learning applications. It provides pre-built installation packages (wheels) so you can quickly get started with cutting-edge AI research and development without complex setup. AI developers, researchers, and machine learning engineers working with PyTorch and CUDA will find this useful.

Use this if you are a developer or researcher who needs to install Flash Attention 3 easily and quickly on various operating systems, including Windows and Arm, for your deep learning projects.

Not ideal if you are not working with deep learning models or do not require the specific performance optimizations provided by Flash Attention 3.

deep-learning-development pytorch-ecosystem gpu-computing ml-infrastructure ai-model-optimization
No Package No Dependents
Maintenance 10 / 25
Adoption 8 / 25
Maturity 15 / 25
Community 3 / 25

How are scores calculated?

Stars

65

Forks

1

Language

Python

License

Apache-2.0

Last pushed

Mar 04, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/windreamer/flash-attention3-wheels"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.