sbhjt-gr/InferrLM
On-device AI for iOS & Android
Runs GGUF models via llama.cpp and Apple MLX locally, while supporting cloud APIs (OpenAI, Gemini, Anthropic) with custom endpoints. Includes built-in HTTP server exposing REST APIs for network access, multimodal vision support with camera integration, and RAG with local OCR-based document ingestion. Built with Expo/React Native and offers branching conversations, streaming responses, and HuggingFace model downloads.
Stars
66
Forks
13
Language
TypeScript
License
AGPL-3.0
Category
Last pushed
Mar 11, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/sbhjt-gr/InferrLM"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
lfnovo/esperanto
A unified interface for various AI model providers
apocas/restai
RESTai is an AIaaS (AI as a Service) open-source platform. Built on top of LlamaIndex &...
baidubce/bce-qianfan-sdk
Provide best practices for LMOps, as well as elegant and convenient access to the features of...
solygambas/python-openai-projects
13 projects using ChatGPT API, Whisper, Embeddings, and DALL-E with Python.
maragudk/gai
Go Artificial Intelligence (GAI) helps you work with foundational models, large language models,...