SergiuDeveloper/yoro-finetuning
YORO (You-Only-Reason-Once) - a novel LLM architecture that runs the main reasoning block once, caches its output, and reuses it for all subsequent tokens. Lightweight auxiliary networks compensate for the missing reasoning passes, keeping generation coherent while skipping the most expensive computation at every step.
Stars
—
Forks
—
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Mar 18, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/SergiuDeveloper/yoro-finetuning"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
OptimalScale/LMFlow
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
adithya-s-k/AI-Engineering.academy
Mastering Applied AI, One Concept at a Time
jax-ml/jax-llm-examples
Minimal yet performant LLM examples in pure JAX
young-geng/scalax
A simple library for scaling up JAX programs
riyanshibohra/TuneKit
Upload your data → Get a fine-tuned SLM. Free.