ProbioticFarmer/mlx-deterministic

Batch-invariant operations for deterministic LLM inference on Apple Silicon using MLX

38
/ 100
Emerging

When performing large language model (LLM) inference on Apple Silicon, you might notice that the same prompt yields slightly different responses depending on how many prompts you process at once (batch size). This tool ensures that your LLM outputs are always identical and reproducible, regardless of the batch size. It provides specific operations that maintain output consistency, making your LLM-powered applications more reliable. This is for AI/ML practitioners and researchers who need consistent and verifiable LLM outputs.

Use this if you need to guarantee that your LLM generates bitwise-identical outputs for the same input, regardless of the batch size used, which is critical for testing, validation, and auditing.

Not ideal if your primary concern is raw inference speed and you can tolerate minor variations in LLM outputs between different batch sizes.

LLM-inference AI-reproducibility ML-validation Model-testing Responsible-AI
No Package No Dependents
Maintenance 6 / 25
Adoption 4 / 25
Maturity 15 / 25
Community 13 / 25

How are scores calculated?

Stars

7

Forks

2

Language

Python

License

MIT

Last pushed

Dec 12, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/ProbioticFarmer/mlx-deterministic"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.