verifai/multiLLM

🚀 Invoke multiple large language models concurrently and the rank results. Add new models and ranking functions quickly.. ⭐

46
/ 100
Emerging

This tool helps technical professionals like engineers or researchers compare responses from different large language models (LLMs) side-by-side. You provide a prompt and API keys for the LLMs you want to use, and it returns each model's answer, along with an explanation of which response is better and why. This is useful for quickly evaluating and choosing the best LLM for a specific coding task, content generation, or problem-solving.

No commits in the last 6 months. Available on PyPI.

Use this if you need to quickly get and compare responses from multiple large language models for a single prompt and want an automated ranking of the best answer.

Not ideal if you need a complex, production-ready system for routing requests to LLMs or for sophisticated A/B testing with detailed metrics.

AI evaluation prompt engineering code generation content comparison technical problem-solving
Stale 6m No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 25 / 25
Community 12 / 25

How are scores calculated?

Stars

83

Forks

9

Language

Python

License

MIT

Last pushed

Nov 06, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/verifai/multiLLM"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.