sjnims/cc-plugin-eval

4-stage evaluation framework for testing Claude Code plugin component triggering. Validates skills, agents, and commands activate correctly via programmatic detection and LLM judgment.

28
/ 100
Experimental

This tool helps AI developers automatically test if their Claude Code plugin components trigger correctly based on user prompts. It takes your plugin's code as input and generates a comprehensive report indicating whether skills, agents, and commands activate as expected or unexpectedly. This is for AI developers, prompt engineers, and QA specialists who build and maintain Claude Code plugins.

Use this if you need to reliably verify that your Claude Code plugin's components (skills, agents, commands) activate precisely when intended by diverse user prompts.

Not ideal if you are testing general LLM conversational abilities or user interfaces outside of Claude Code plugin component triggering.

AI development LLM plugin testing Claude Code Prompt engineering Software quality assurance
No Package No Dependents
Maintenance 10 / 25
Adoption 5 / 25
Maturity 13 / 25
Community 0 / 25

How are scores calculated?

Stars

14

Forks

Language

TypeScript

License

MIT

Last pushed

Mar 10, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/agents/sjnims/cc-plugin-eval"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.