lechmazur/writing
This benchmark tests how well LLMs incorporate a set of 10 mandatory story elements (characters, objects, core concepts, attributes, motivations, etc.) in a short creative story
This benchmark helps creative writers and content strategists understand which large language models (LLMs) are best at writing engaging short stories while precisely following specific creative instructions. It takes a creative brief with ten required story elements as input and produces a scorecard showing how well different LLMs integrated those elements and achieved literary quality. Anyone tasked with generating high-quality creative content using AI would find this useful.
353 stars.
Use this if you need to choose an LLM for creative writing tasks where adherence to a detailed brief and narrative quality are both critical.
Not ideal if your primary goal is generating factual reports, code, or non-narrative content, or if you need to compare LLMs on speed or cost.
Stars
353
Forks
8
Language
Batchfile
License
—
Category
Last pushed
Feb 06, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/lechmazur/writing"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
TsinghuaC3I/MARTI
A Framework for LLM-based Multi-Agent Reinforced Training and Inference
zjunlp/KnowLM
An Open-sourced Knowledgable Large Language Model Framework.
cli99/llm-analysis
Latency and Memory Analysis of Transformer Models for Training and Inference
tanyuqian/redco
NAACL '24 (Best Demo Paper RunnerUp) / MlSys @ NeurIPS '23 - RedCoast: A Lightweight Tool to...
stanleylsx/llms_tool
一个基于HuggingFace开发的大语言模型训练、测试工具。支持各模型的webui、终端预测,低参数量及全参数模型训练(预训练、SFT、RM、PPO、DPO)和融合、量化。