parea-ai/parea-sdk-py
Python SDK for experimenting, testing, evaluating & monitoring LLM-powered applications - Parea AI (YC S23)
This tool helps AI application developers test, evaluate, and monitor their Large Language Model (LLM) applications. It takes your LLM code and test datasets as input, providing performance scores, detailed logs, and a dashboard view of your application's behavior. Developers building with LLMs will find this useful for ensuring their AI models perform as expected.
No commits in the last 6 months.
Use this if you are developing LLM-powered applications and need a systematic way to test performance, observe how your models process information, and collaborate on prompt engineering.
Not ideal if you are looking for a general-purpose machine learning model evaluation tool not specifically focused on LLMs.
Stars
82
Forks
9
Language
Python
License
Apache-2.0
Category
Last pushed
Feb 13, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/parea-ai/parea-sdk-py"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Mirascope/lilypad
Open-source versioning, tracing, and annotation tooling.
Supervertaler/Supervertaler-Workbench
Open-source, AI-enhanced CAT tool with multi-LLM support, translation memory, glossary...
crjaensch/PromptoLab
A multi-platform app to serve as a prompts catalog, a LLM playground for running and optimizing...
geeknees/sentinel_rb
SentinelRb is an LLM-driven prompt inspector designed to automatically detect common...
NeuroTinkerLab/synt-e-project
A Python tool to translate natural language requests into efficient, single-line commands for AI...