campfirein/brv-bench

Benchmark suite for evaluating retrieval quality and latency of AI agent context systems

30
/ 100
Emerging

This tool helps AI application developers measure how well their AI agents retrieve context for conversations. You provide existing conversational datasets, and it evaluates the AI agent's ability to recall relevant information, measuring accuracy and speed. This is ideal for AI engineers, product managers, or researchers building and refining conversational AI applications.

Use this if you need to objectively assess the quality and performance of your AI agent's long-term conversational memory and context retrieval capabilities.

Not ideal if you are looking for a general-purpose AI agent development framework or a tool to build AI agents from scratch.

AI Agent Development Conversational AI Retrieval Augmented Generation AI Performance Testing Large Language Models
No License No Package No Dependents
Maintenance 10 / 25
Adoption 5 / 25
Maturity 3 / 25
Community 12 / 25

How are scores calculated?

Stars

11

Forks

2

Language

Python

License

Last pushed

Mar 12, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/agents/campfirein/brv-bench"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.