thkox/home-ai-server
Home AI Server provides the backend infrastructure for the Home AI system. It manages user data, conversations, and the interaction with Large Language Models (LLMs). The server runs locally, ensuring that all user data stays within the local network for maximum privacy.
No commits in the last 6 months.
Stars
4
Forks
—
Language
Python
License
MIT
Category
Last pushed
Oct 18, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/thkox/home-ai-server"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
containers/ramalama
RamaLama is an open-source developer tool that simplifies the local serving of AI models from...
av/harbor
One command brings a complete pre-wired LLM stack with hundreds of services to explore.
RunanywhereAI/runanywhere-sdks
Production ready toolkit to run AI locally
runpod-workers/worker-vllm
The RunPod worker template for serving our large language model endpoints. Powered by vLLM.
vtuber-plan/olah
Self-hosted huggingface mirror service. 自建huggingface镜像服务。