Reason-Wang/NAT

[NAACL 2025] The official implementation of paper "Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language Models as Agents"

15
/ 100
Experimental

This project helps improve the reasoning capabilities of large language models (LLMs) used as AI agents. By integrating both successful and unsuccessful attempts at problem-solving, it teaches LLMs to avoid common pitfalls. The input is a collection of problem-solving attempts, and the output is a more robust LLM agent. This is for researchers and practitioners who are fine-tuning LLMs for complex tasks.

No commits in the last 6 months.

Use this if you are fine-tuning large language models to act as intelligent agents and want them to perform better on mathematical reasoning or complex question-answering tasks by learning from mistakes.

Not ideal if you are looking for a pre-trained general-purpose LLM without specific agentic fine-tuning needs or if your primary focus is on generative text rather than problem-solving.

AI agent training LLM fine-tuning mathematical reasoning question answering natural language understanding
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 8 / 25
Community 0 / 25

How are scores calculated?

Stars

28

Forks

Language

Python

License

Last pushed

Mar 14, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Reason-Wang/NAT"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.