equinor/promptly

A prompt collection for testing and evaluation of LLMs.

43
/ 100
Emerging

This collection provides pre-written prompts for evaluating and testing Large Language Models (LLMs). It helps you put different LLMs through their paces, feeding them specific questions and scenarios to assess their performance and responses. Scientific programmers and researchers who work with AI models would find this useful for benchmarking and understanding LLM capabilities.

Use this if you need a structured set of prompts to systematically test and compare various Large Language Models.

Not ideal if you are looking for a tool to generate prompts automatically or for general-purpose prompt engineering outside of evaluation.

LLM-evaluation AI-testing prompt-benchmarking AI-research natural-language-processing
No Package No Dependents
Maintenance 10 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 10 / 25

How are scores calculated?

Stars

27

Forks

3

Language

Jupyter Notebook

License

CC-BY-4.0

Last pushed

Feb 24, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/equinor/promptly"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.