RunanywhereAI/RCLI

Talk to your Mac, query your docs, no cloud required. On-device voice AI + RAG

60
/ 100
Established

Implements a complete STT + LLM + TTS + VLM pipeline optimized for Apple Silicon via MetalRT, a proprietary GPU inference engine delivering sub-200ms latency and 714x faster speech recognition than real-time. Includes hybrid vector+BM25 RAG (~4ms retrieval), 40 native macOS voice actions (Spotify control, app launching, system commands), and on-device vision capabilities for camera/screen analysis without model quantization. Supports hot-swappable models (Qwen3, LFM2, Whisper) with graceful fallback to llama.cpp for M1/M2 Macs.

1,054 stars. Actively maintained with 89 commits in the last 30 days.

No Package No Dependents
Maintenance 25 / 25
Adoption 10 / 25
Maturity 11 / 25
Community 14 / 25

How are scores calculated?

Stars

1,054

Forks

47

Language

C++

License

MIT

Last pushed

Mar 13, 2026

Commits (30d)

89

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/rag/RunanywhereAI/RCLI"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.