stashlabs/duelr

Compare LLMs in one click

32
/ 100
Emerging

This tool helps AI application developers quickly compare how different large language models (LLMs) respond to specific prompts. You input a prompt and select multiple LLM models, then receive a side-by-side comparison of their responses along with metrics like speed, cost, and output quality. It's for anyone building or integrating LLMs who needs to choose the best model for a task.

No commits in the last 6 months.

Use this if you need to evaluate multiple LLMs for a specific use case, comparing their performance on factors like response quality, speed, and cost.

Not ideal if you're looking for a fully managed service or don't have the technical comfort to install and run a local application with API keys.

AI-application-development LLM-evaluation prompt-engineering AI-model-selection
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 7 / 25
Maturity 15 / 25
Community 8 / 25

How are scores calculated?

Stars

39

Forks

3

Language

TypeScript

License

Apache-2.0

Last pushed

Aug 08, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/stashlabs/duelr"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.