zhenyi4/codi

Official repository for "CODI: Compressing Chain-of-Thought into Continuous Space via Self-Distillation"

39
/ 100
Emerging

This project offers a way to distill the complex problem-solving steps of large language models (LLMs) into a more efficient, compressed format. It takes a problem and the LLM's step-by-step reasoning (known as Chain-of-Thought) and produces a smaller, faster model that can achieve similar reasoning capabilities. This is for researchers and engineers working with LLMs who need to make their models more performant and less resource-intensive.

Use this if you are a machine learning researcher or engineer aiming to compress the reasoning capabilities of large language models for efficiency without significant performance loss.

Not ideal if you are an end-user simply looking to apply an existing language model for text generation or analysis without modifying its core architecture or training process.

large-language-models model-compression natural-language-processing machine-learning-research ai-efficiency
No License No Package No Dependents
Maintenance 6 / 25
Adoption 9 / 25
Maturity 7 / 25
Community 17 / 25

How are scores calculated?

Stars

73

Forks

13

Language

Python

License

Last pushed

Dec 15, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/zhenyi4/codi"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.