fau-masters-collected-works-cgarbin/gpt-all-local
A "chat with your data" example: using a large language models (LLM) to interact with our own (local) data. Everything is local: the embedding model, the LLM, the vector database. This is an example of retrieval-augmented generation (RAG): we find relevant sections from our documents and pass it to the LLM as part of the prompt (see pics).
Stars
30
Forks
5
Language
Python
License
—
Category
Last pushed
Jan 15, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/vector-db/fau-masters-collected-works-cgarbin/gpt-all-local"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
biocypher/biochatter
Backend library for conversational AI in biomedicine
7-docs/7-docs
Use local files or public GitHub repository as a source and ask questions through ChatGPT about it
pgalko/BambooAI
A Python library powered by Language Models (LLMs) for conversational data discovery and analysis.
redis-developer/ArXivChatGuru
Use ArXiv ChatGuru to talk to research papers. This app uses LangChain, OpenAI, Streamlit, and...
redis-developer/LLM-Document-Chat
Using LlamaIndex, Redis, and OpenAI to chat with PDF documents. Supplementary material for blog...