LLM-Hub and ToolNeuron
These are competitors offering overlapping local LLM functionality on Android devices, with ToolNeuron providing a more comprehensive feature set (VLM, image generation, RAG, TTS/STT) compared to LLM-Hub's narrower chat-focused approach.
About LLM-Hub
timmyy123/LLM-Hub
Local AI Assistant on your phone
Enables on-device multi-modal AI with chat, image generation, code completion, and specialized tools (transcription, translation, OCR-based scam detection) using optimized model formats (LiteRT Task, GGUF, ONNX) that leverage GPU/NPU acceleration across multiple model families. Built with Kotlin/Jetpack Compose, it supports flexible model imports from HuggingFace and includes Kid Mode with model-level safety guardrails, RAG memory, and web search integration—all with zero data collection and 100% local inference.
About ToolNeuron
Siddhesh2377/ToolNeuron
On-device AI for Android — LLM chat (GGUF/llama.cpp), vision models (VLM), image generation (Stable Diffusion), tool calling, AI personas, RAG knowledge packs, TTS/STT. Fully offline, zero subscriptions, open-source.
Implements a hybrid RAG pipeline combining FTS4 BM25, vector search, and Reciprocal Rank Fusion across PDFs, docs, and chat history, with grammar-constrained JSON for function calling. Built on llama.cpp for inference, LocalDream for Stable Diffusion, and a custom Unified Memory System (UMS) for efficient binary storage, with AES-256-GCM encryption backed by Android KeyStore and hardware TEE where available. Includes 7 callable plugins (web search, file ops, calculator) and adaptive hardware tuning that auto-detects CPU topology to optimize thread allocation and context windows.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work