HIDEKI-SQ/cot-backfire
When does chain-of-thought prompting backfire? Experiments identifying the critical alignment threshold for provided reasoning traces.
Stars
—
Forks
—
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Feb 21, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/HIDEKI-SQ/cot-backfire"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
kyegomez/tree-of-thoughts
Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large...
spcl/graph-of-thoughts
Official Implementation of "Graph of Thoughts: Solving Elaborate Problems with Large Language Models"
kyegomez/the-compiler
Seed, Code, Harvest: Grow Your Own App with Tree of Thoughts!
habedi/cogitator
A Python toolkit for chain-of-thought prompting 🐍
amazon-science/auto-cot
Official implementation for "Automatic Chain of Thought Prompting in Large Language Models"...