Llm Comparison Evaluation AI Coding Tools

There are 7 llm comparison evaluation tools tracked. The highest-rated is THU-WingTecher/LSPRAG at 46/100 with 33 stars.

Get all 7 projects as JSON

curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=ai-coding&subcategory=llm-comparison-evaluation&limit=20"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.

# Tool Score Tier
1 THU-WingTecher/LSPRAG

Real-time multi-language unit test generation tool via LSP

46
Emerging
2 metareflection/dafny-replay

Verified kernels, written in Dafny and compiled to JavaScript, for...

34
Emerging
3 santinic/unvibe

Generate correct code from unit-tests

33
Emerging
4 bigcode-project/bigcodearena

BigCodeArena: Unveiling More Reliable Human Preferences in Code Generation...

32
Emerging
5 adilanwar2399/ESBMC-ibmc

The ESBMC ibmc (Invariant Based Model Checking) Tool.

27
Experimental
6 mpuodziukas-labs/cobol-demo

COBOL modernization: LLMs introduce bugs, humans validate. Production-grade...

14
Experimental
7 LeonardoSaccotelli/Formal-Methods-For-Computer-Science

Collection of different tasks implemented in the context of Formal Methods...

11
Experimental