CS-EVAL/CS-Eval

CS-Eval is a comprehensive evaluation suite for fundamental cybersecurity models or large language models' cybersecurity ability.

35
/ 100
Emerging

This toolkit helps cybersecurity professionals and AI developers assess the cybersecurity knowledge and reasoning abilities of large language models (LLMs). You input a cybersecurity model or an LLM, and it provides an evaluation report across 11 major cybersecurity categories and 42 subdomains. This allows you to understand a model's strengths and weaknesses in cybersecurity.

No commits in the last 6 months.

Use this if you need to objectively benchmark and compare the cybersecurity capabilities of different AI models or LLMs for applications like threat intelligence, security operations, or vulnerability analysis.

Not ideal if you are looking for a tool to secure your own systems or detect live threats; this is an evaluation tool, not a security solution.

cybersecurity-evaluation AI-security model-benchmarking threat-intelligence security-operations
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

60

Forks

6

Language

License

MIT

Last pushed

Nov 27, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/CS-EVAL/CS-Eval"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.