laiso/ts-bench
Measure and compare the performance of AI coding agents on TypeScript tasks.
This tool helps AI developers and researchers evaluate the performance of different AI coding agents on TypeScript programming tasks. You provide the agent (like Claude) and a TypeScript coding challenge, and it outputs pass/fail results to compare how well various models generate correct code. This is for anyone creating, fine-tuning, or selecting AI models for code generation.
210 stars.
Use this if you need to quickly benchmark and compare the effectiveness of various AI coding agents in solving TypeScript programming problems.
Not ideal if you require lab-grade, highly precise, and statistically rigorous performance evaluations of AI models.
Stars
210
Forks
10
Language
TypeScript
License
—
Category
Last pushed
Mar 12, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/agents/laiso/ts-bench"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Higher-rated alternatives
StonyBrookNLP/appworld
🌍 AppWorld: A Controllable World of Apps and People for Benchmarking Function Calling and...
qualifire-dev/rogue
AI Agent Evaluator & Red Team Platform
microsoft/WindowsAgentArena
Windows Agent Arena (WAA) 🪟 is a scalable OS platform for testing and benchmarking of...
future-agi/ai-evaluation
Evaluation Framework for all your AI related Workflows
agentscope-ai/OpenJudge
OpenJudge: A Unified Framework for Holistic Evaluation and Quality Rewards