RunanywhereAI/RCLI
Talk to your Mac, query your docs, no cloud required. On-device voice AI + RAG
Implements a complete STT + LLM + TTS + VLM pipeline optimized for Apple Silicon via MetalRT, a proprietary GPU inference engine delivering sub-200ms latency and 714x faster speech recognition than real-time. Includes hybrid vector+BM25 RAG (~4ms retrieval), 40 native macOS voice actions (Spotify control, app launching, system commands), and on-device vision capabilities for camera/screen analysis without model quantization. Supports hot-swappable models (Qwen3, LFM2, Whisper) with graceful fallback to llama.cpp for M1/M2 Macs.
1,054 stars. Actively maintained with 89 commits in the last 30 days.
Stars
1,054
Forks
47
Language
C++
License
MIT
Category
Last pushed
Mar 13, 2026
Commits (30d)
89
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/rag/RunanywhereAI/RCLI"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
labring/FastGPT
FastGPT is a knowledge-based platform built on the LLMs, offers a comprehensive suite of...
ragpi/ragpi
🤖 An open-source AI assistant answering questions using your docs
theaiautomators/insights-lm-local-package
Open-source, fully private and local alternative to NotebookLM. Chat with your documents,...
AstraBert/PapersChat
An agentic AI application that allows you to chat with your papers and gather also information...
isala404/Tera
Tera is an AI assistant which is tailored just for you and runs fully locally.