LLM-Hub and ToolNeuron

These are competitors offering overlapping local LLM functionality on Android devices, with ToolNeuron providing a more comprehensive feature set (VLM, image generation, RAG, TTS/STT) compared to LLM-Hub's narrower chat-focused approach.

LLM-Hub
60
Established
ToolNeuron
51
Established
Maintenance 13/25
Adoption 10/25
Maturity 15/25
Community 22/25
Maintenance 13/25
Adoption 10/25
Maturity 15/25
Community 13/25
Stars: 196
Forks: 49
Downloads:
Commits (30d): 0
Language: Kotlin
License:
Stars: 272
Forks: 20
Downloads:
Commits (30d): 0
Language: Kotlin
License: Apache-2.0
No Package No Dependents
No Package No Dependents

About LLM-Hub

timmyy123/LLM-Hub

Local AI Assistant on your phone

Enables on-device multi-modal AI with chat, image generation, code completion, and specialized tools (transcription, translation, OCR-based scam detection) using optimized model formats (LiteRT Task, GGUF, ONNX) that leverage GPU/NPU acceleration across multiple model families. Built with Kotlin/Jetpack Compose, it supports flexible model imports from HuggingFace and includes Kid Mode with model-level safety guardrails, RAG memory, and web search integration—all with zero data collection and 100% local inference.

About ToolNeuron

Siddhesh2377/ToolNeuron

On-device AI for Android — LLM chat (GGUF/llama.cpp), vision models (VLM), image generation (Stable Diffusion), tool calling, AI personas, RAG knowledge packs, TTS/STT. Fully offline, zero subscriptions, open-source.

Implements a hybrid RAG pipeline combining FTS4 BM25, vector search, and Reciprocal Rank Fusion across PDFs, docs, and chat history, with grammar-constrained JSON for function calling. Built on llama.cpp for inference, LocalDream for Stable Diffusion, and a custom Unified Memory System (UMS) for efficient binary storage, with AES-256-GCM encryption backed by Android KeyStore and hardware TEE where available. Includes 7 callable plugins (web search, file ops, calculator) and adaptive hardware tuning that auto-detects CPU topology to optimize thread allocation and context windows.

Related comparisons

Scores updated daily from GitHub, PyPI, and npm data. How scores work