jackso1328/code-review-benchmark-env
A production-grade OpenEnv benchmark for AI code reviewers. Features multi-file reasoning, shaped rewards, and security-focused tasks (SQL Injection, O(n²) complexity).
Stars
—
Forks
—
Language
Python
License
MIT
Category
Last pushed
Apr 04, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/agents/jackso1328/code-review-benchmark-env"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
alvinreal/awesome-autoresearch
A curated list of autonomous improvement loops, research agents, and autoresearch-style systems...
wulinteousa2-hash/napari-chat-assistant
A local agent architecture for semantic-aware interaction between large language models and...
0xSteph/pentest-ai-agents
Turn Claude Code into your offensive security research assistant. Specialized AI subagents for...
saksham-jain177/AI-Agent-based-Deep-Research
Deep Research AI Agent is a dual-agent system that conducts web-based research and generates...
theam/limina
Autonomous research harness for AI agents. Give it a measurable goal — it hypothesizes,...