aws-samples/rag-with-amazon-opensearch-and-sagemaker
Question Answering Generative AI application with Large Language Models (LLMs) and Amazon OpenSearch Service
Implements retrieval-augmented generation by storing document embeddings in OpenSearch and dynamically retrieving relevant passages to augment LLM prompts, addressing token limits and improving answer accuracy. Deploys SageMaker endpoints for both text generation and embedding creation, with infrastructure-as-code (CDK) for the full stack including OpenSearch clusters and credential management. Provides a complete end-to-end workflow from data ingestion through a Streamlit frontend, leveraging LangChain for orchestration.
No commits in the last 6 months.
Stars
29
Forks
3
Language
Python
License
MIT-0
Last pushed
Dec 03, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/vector-db/aws-samples/rag-with-amazon-opensearch-and-sagemaker"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Higher-rated alternatives
Siddhant-K-code/distill
Reliable LLM outputs start with clean context. Deterministic deduplication, compression, and...
louisbrulenaudet/ragoon
High level library for batched embeddings generation, blazingly-fast web-based RAG and quantized...
pesu-dev/ask-pesu
A RAG pipeline for question answering about PES University
namtroi/RAGBase
Open Source RAG ETL Platform. Turns PDFs, Docs & Slides into queryable vectors. Features a...
B-A-M-N/FlockParser
Distributed document RAG system with intelligent GPU/CPU orchestration. Auto-discovers...