Sparse Attention Optimization LLM Tools
Efficient sparse attention implementations and frameworks that reduce computational complexity for LLM inference and training. Includes kernel optimizations, attention pattern strategies, and performance-tuned libraries. Does NOT include general attention mechanisms, model architectures, or non-attention-specific optimization techniques.
There are 2 sparse attention optimization tools tracked. The highest-rated is windreamer/flash-attention3-wheels at 36/100 with 65 stars.
Get all 2 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=llm-tools&subcategory=sparse-attention-optimization&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Tool | Score | Tier |
|---|---|---|---|
| 1 |
windreamer/flash-attention3-wheels
Pre-built wheels that erase Flash Attention 3 installation headaches. |
|
Emerging |
| 2 |
etasnadi/VulkanCooperativeMatrixAttention
Vulkan & GLSL implementation of FlashAttention-2 |
|
Experimental |