Ziechoes/reasoning-invariance-benchmark
Experiments testing whether LLM reasoning trajectories remain invariant when constraint layers are applied. If reasoning paths diverge under identical logical problems, this suggests architectural coupling between inference state and constraint enforcement.
Stars
1
Forks
—
Language
Python
License
—
Category
Last pushed
Mar 04, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/Ziechoes/reasoning-invariance-benchmark"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Higher-rated alternatives
microsoft/promptbench
A unified evaluation framework for large language models
uptrain-ai/uptrain
UpTrain is an open-source unified platform to evaluate and improve Generative AI applications....
gabe-mousa/Apolien
AI Safety Evaluation Library
microsoftarchive/promptbench
A unified evaluation framework for large language models
babelcloud/LLM-RGB
LLM Reasoning and Generation Benchmark. Evaluate LLMs in complex scenarios systematically.