marklysze/AutoGenPromptTesting

Prompt testing Local LLMs for Microsoft's AutoGen

20
/ 100
Experimental

This project helps developers evaluate how well different local Large Language Models (LLMs) perform specific tasks within an AutoGen multi-agent system. It takes various local LLMs and different prompt engineering techniques as input, then measures how accurately the LLM selects the correct 'speaker' or agent in a simulated debate scenario. The ideal user is a developer building multi-agent AI applications with AutoGen, specifically when using local, self-hosted LLMs that might not be as capable as commercial alternatives.

No commits in the last 6 months.

Use this if you are an AI application developer working with AutoGen and need to understand how to optimize prompts for local LLMs to achieve reliable multi-agent coordination.

Not ideal if you are looking for a fully structured, production-ready framework for generalized prompt testing, or if you are exclusively using highly capable private LLMs like ChatGPT.

AI application development AutoGen local LLMs prompt engineering multi-agent systems
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 8 / 25
Community 7 / 25

How are scores calculated?

Stars

10

Forks

1

Language

Python

License

Last pushed

Feb 27, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/marklysze/AutoGenPromptTesting"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.