hao-ai-lab/Consistency_LLM
[ICML 2024] CLLMs: Consistency Large Language Models
This project offers a way to make large language models (LLMs) generate text, code, or answers much faster. It takes an existing LLM, along with a dataset you've used to train it, and outputs a specialized version of that LLM that can produce results several times quicker. This is ideal for developers or researchers who build and deploy LLMs for applications like chatbots, code generation tools, or problem-solving assistants.
413 stars. No commits in the last 6 months.
Use this if you are a developer or researcher looking to significantly speed up the inference (response time) of your existing large language models without complex architectural changes or needing a 'draft' model.
Not ideal if you are looking for a pre-built, end-user application or if you do not have the technical expertise to work with and train large language models.
Stars
413
Forks
22
Language
Python
License
Apache-2.0
Category
Last pushed
Nov 16, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/hao-ai-lab/Consistency_LLM"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
MadryLab/context-cite
Attribute (or cite) statements generated by LLMs back to in-context information.
microsoft/augmented-interpretable-models
Interpretable and efficient predictors using pre-trained language models. Scikit-learn compatible.
Trustworthy-ML-Lab/CB-LLMs
[ICLR 25] A novel framework for building intrinsically interpretable LLMs with...
poloclub/LLM-Attributor
LLM Attributor: Attribute LLM's Generated Text to Training Data
THUDM/LongCite
LongCite: Enabling LLMs to Generate Fine-grained Citations in Long-context QA