Llm Comparison Evaluation AI Coding Tools
There are 7 llm comparison evaluation tools tracked. The highest-rated is THU-WingTecher/LSPRAG at 46/100 with 33 stars.
Get all 7 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=ai-coding&subcategory=llm-comparison-evaluation&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Tool | Score | Tier |
|---|---|---|---|
| 1 |
THU-WingTecher/LSPRAG
Real-time multi-language unit test generation tool via LSP |
|
Emerging |
| 2 |
metareflection/dafny-replay
Verified kernels, written in Dafny and compiled to JavaScript, for... |
|
Emerging |
| 3 |
santinic/unvibe
Generate correct code from unit-tests |
|
Emerging |
| 4 |
bigcode-project/bigcodearena
BigCodeArena: Unveiling More Reliable Human Preferences in Code Generation... |
|
Emerging |
| 5 |
adilanwar2399/ESBMC-ibmc
The ESBMC ibmc (Invariant Based Model Checking) Tool. |
|
Experimental |
| 6 |
mpuodziukas-labs/cobol-demo
COBOL modernization: LLMs introduce bugs, humans validate. Production-grade... |
|
Experimental |
| 7 |
LeonardoSaccotelli/Formal-Methods-For-Computer-Science
Collection of different tasks implemented in the context of Formal Methods... |
|
Experimental |