Marker-Inc-Korea/AutoRAG
AutoRAG: An Open-Source Framework for Retrieval-Augmented Generation (RAG) Evaluation & Optimization with AutoML-Style Automation
Provides end-to-end RAG pipeline optimization through YAML-driven configuration, encompassing document parsing, semantic chunking, and QA dataset generation with support for multiple parsing/chunking strategies simultaneously. Uses grid-search and metric-driven evaluation across retriever-generator combinations to identify optimal module configurations, with results tracked in a dashboard for deployment-ready pipeline export. Integrates with LlamaIndex, LangChain, and local embedding models, supporting both cloud APIs (OpenAI) and GPU-accelerated inference for custom models.
4,609 stars. Actively maintained with 5 commits in the last 30 days. Available on PyPI.
Stars
4,609
Forks
381
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 10, 2026
Commits (30d)
5
Dependencies
59
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/rag/Marker-Inc-Korea/AutoRAG"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related tools
IntelLabs/RAG-FiT
Framework for enhancing LLMs for RAG tasks using fine-tuning.
jxzhangjhu/Awesome-LLM-RAG
Awesome-LLM-RAG: a curated list of advanced retrieval augmented generation (RAG) in Large Language Models
coree/awesome-rag
A curated list of retrieval-augmented generation (RAG) in large language models
IntelLabs/fastRAG
Efficient Retrieval Augmentation and Generation Framework
ibm-self-serve-assets/Blended-RAG
Blended RAG: Improving RAG (Retriever-Augmented Generation) Accuracy with Semantic Search and...