Ranjitbarnala0/rai
CPU-native LLM inference engine — hand-written SIMD kernels, 4-bit quantized, zero GPU required. Pure Rust.
Stars
—
Forks
—
Language
Rust
License
MIT
Category
Last pushed
Mar 12, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Ranjitbarnala0/rai"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
EricLBuehler/mistral.rs
Fast, flexible LLM inference
nerdai/llms-from-scratch-rs
A comprehensive Rust translation of the code from Sebastian Raschka's Build an LLM from Scratch book.
brontoguana/krasis
Krasis is a Hybrid LLM runtime which focuses on efficient running of larger models on consumer...
ShelbyJenkins/llm_utils
llm_utils: Basic LLM tools, best practices, and minimal abstraction.
Mattbusel/llm-wasm
LLM inference primitives for WebAssembly — cache, retry, routing, guards, cost tracking, templates