weizhepei/InstructRAG
[ICLR 2025] InstructRAG: Instructing Retrieval-Augmented Generation via Self-Synthesized Rationales
Employs instruction-tuned LMs to generate self-supervised rationales that explicitly denoise retrieved passages, improving verifiability without external labels. Supports both in-context learning and supervised fine-tuning workflows across multiple QA benchmarks (PopQA, TriviaQA, Natural Questions, ASQA, 2WikiMultiHopQA) with pluggable retrievers like Contriever, DPR, and BM25. Pre-trained checkpoints and training scripts target multi-GPU setups, enabling 8%+ performance gains and robustness to noisy retrieval contexts.
138 stars. No commits in the last 6 months.
Stars
138
Forks
9
Language
Python
License
MIT
Category
Last pushed
Feb 06, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/rag/weizhepei/InstructRAG"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Renumics/renumics-rag
Visualization for a Retrieval-Augmented Generation (RAG) Assistant 🤖❤️📚
VectorInstitute/retrieval-augmented-generation
Reference Implementations for the RAG bootcamp
naver/bergen
Benchmarking library for RAG
KalyanKS-NLP/rag-zero-to-hero-guide
Comprehensive guide to learn RAG from basics to advanced.
alan-turing-institute/t0-1
Application of Retrieval-Augmented Reasoning on a domain-specific body of knowledge