lechmazur/writing

This benchmark tests how well LLMs incorporate a set of 10 mandatory story elements (characters, objects, core concepts, attributes, motivations, etc.) in a short creative story

36
/ 100
Emerging

This benchmark helps creative writers and content strategists understand which large language models (LLMs) are best at writing engaging short stories while precisely following specific creative instructions. It takes a creative brief with ten required story elements as input and produces a scorecard showing how well different LLMs integrated those elements and achieved literary quality. Anyone tasked with generating high-quality creative content using AI would find this useful.

353 stars.

Use this if you need to choose an LLM for creative writing tasks where adherence to a detailed brief and narrative quality are both critical.

Not ideal if your primary goal is generating factual reports, code, or non-narrative content, or if you need to compare LLMs on speed or cost.

creative-writing content-creation narrative-design AI-evaluation storytelling
No License No Package No Dependents
Maintenance 10 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 8 / 25

How are scores calculated?

Stars

353

Forks

8

Language

Batchfile

License

Last pushed

Feb 06, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/lechmazur/writing"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.