tommasocerruti/detllm
Deterministic-mode checks for LLM inference: measure run/batch variance, generate repro packs, and explain why outputs differ.
When you're working with AI models that generate text, you might notice the same prompt can sometimes give different answers. This tool helps you understand why by taking your text prompts and an AI model, then showing you if the outputs vary across different runs or batch sizes. It's for AI engineers, data scientists, or anyone developing or testing large language models who needs consistent and predictable results.
Use this if you need to ensure that your language model consistently produces the same output for the same input, or if you need to diagnose why it isn't.
Not ideal if you are troubleshooting determinism issues in distributed or multi-process AI inference environments, or if you don't need strict reproducibility for your model outputs.
Stars
18
Forks
1
Language
Python
License
Apache-2.0
Category
Last pushed
Jan 17, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/tommasocerruti/detllm"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
sgl-project/sglang
SGLang is a high-performance serving framework for large language models and multimodal models.
alibaba/MNN
MNN: A blazing-fast, lightweight inference engine battle-tested by Alibaba, powering...
xorbitsai/inference
Swap GPT for any LLM by changing a single line of code. Xinference lets you run open-source,...
tensorzero/tensorzero
TensorZero is an open-source stack for industrial-grade LLM applications. It unifies an LLM...