stashlabs/duelr
Compare LLMs in one click
This tool helps AI application developers quickly compare how different large language models (LLMs) respond to specific prompts. You input a prompt and select multiple LLM models, then receive a side-by-side comparison of their responses along with metrics like speed, cost, and output quality. It's for anyone building or integrating LLMs who needs to choose the best model for a task.
No commits in the last 6 months.
Use this if you need to evaluate multiple LLMs for a specific use case, comparing their performance on factors like response quality, speed, and cost.
Not ideal if you're looking for a fully managed service or don't have the technical comfort to install and run a local application with API keys.
Stars
39
Forks
3
Language
TypeScript
License
Apache-2.0
Category
Last pushed
Aug 08, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/stashlabs/duelr"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Higher-rated alternatives
open-compass/opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral,...
IBM/unitxt
🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the...
lean-dojo/LeanDojo
Tool for data extraction and interacting with Lean programmatically.
GoodStartLabs/AI_Diplomacy
Frontier Models playing the board game Diplomacy.
google/litmus
Litmus is a comprehensive LLM testing and evaluation tool designed for GenAI Application...