SpecForge and TorchSpec

SpecForge and TorchSpec are competitors for training speculative decoding models, with SpecForge offering additional integration into the SGLang serving ecosystem while TorchSpec provides a PyTorch-native alternative.

SpecForge
79
Verified
TorchSpec
37
Emerging
Maintenance 20/25
Adoption 10/25
Maturity 24/25
Community 25/25
Maintenance 10/25
Adoption 7/25
Maturity 11/25
Community 9/25
Stars: 729
Forks: 179
Downloads:
Commits (30d): 27
Language: Python
License: MIT
Stars: 32
Forks: 3
Downloads:
Commits (30d): 0
Language: Python
License: MIT
No Dependents
No Package No Dependents

About SpecForge

sgl-project/SpecForge

Train speculative decoding models effortlessly and port them smoothly to SGLang serving.

For those working with large language models, SpecForge helps you train specialized 'speculative decoding' models that can significantly speed up how fast your main LLM responds. You feed in your LLM and it outputs a more efficient version, ready to be used with the SGLang serving framework. This is for AI practitioners and researchers looking to optimize LLM inference performance.

LLM-optimization AI-inference model-training machine-learning-engineering large-language-models

About TorchSpec

torchspec-project/TorchSpec

A PyTorch native library for training speculative decoding models

This tool helps AI engineers optimize large language models by training specialized 'draft' models for speculative decoding. It takes hidden states from existing inference engines as input and produces a smaller, faster draft model. This allows for significant speed improvements when generating text with large language models.

large-language-models LLM-deployment model-optimization AI-inference deep-learning-engineering

Scores updated daily from GitHub, PyPI, and npm data. How scores work