SJTU-DENG-Lab/Discrete-Diffusion-Forcing

Discrete Diffusion Forcing (D2F): dLLMs Can Do Faster-Than-AR Inference

47
/ 100
Emerging

This project helps AI developers and researchers significantly speed up large language model (LLM) inference. It takes an existing diffusion-based LLM and provides a method to generate text, code, or other sequences much faster. The result is quicker responses from LLMs, making them more practical for real-time applications, while maintaining high generation quality.

244 stars.

Use this if you are building or deploying large language models and need to accelerate their text generation speed, especially for high-throughput applications, without sacrificing output quality.

Not ideal if you are working with traditional autoregressive LLMs and are not familiar with diffusion models or advanced model training/inference techniques.

AI-development LLM-deployment inference-optimization natural-language-generation code-generation
No Package No Dependents
Maintenance 10 / 25
Adoption 10 / 25
Maturity 15 / 25
Community 12 / 25

How are scores calculated?

Stars

244

Forks

17

Language

Python

License

MIT

Last pushed

Feb 03, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/SJTU-DENG-Lab/Discrete-Diffusion-Forcing"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.