Llm Comparison Evaluation AI Coding Tools
There are 4 llm comparison evaluation tools tracked. The highest-rated is THU-WingTecher/LSPRAG at 42/100 with 33 stars.
Get all 4 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=ai-coding&subcategory=llm-comparison-evaluation&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Tool | Score | Tier |
|---|---|---|---|
| 1 |
THU-WingTecher/LSPRAG
Real-time multi-language unit test generation tool via LSP |
|
Emerging |
| 2 |
metareflection/dafny-replay
Verified kernels, written in Dafny and compiled to JavaScript, for... |
|
Emerging |
| 3 |
bigcode-project/bigcodearena
BigCodeArena: Unveiling More Reliable Human Preferences in Code Generation... |
|
Emerging |
| 4 |
santinic/unvibe
Generate correct code from unit-tests |
|
Experimental |