Ubaida-M-Yusuf/Makimus-AI
AI-powered media search — find images and videos using natural language or visual queries
Leverages CLIP embeddings via `open_clip_torch` to encode both media and queries into a unified vector space, enabling cross-modal search across images, videos, RAW files, and text. Supports GPU acceleration across NVIDIA (CUDA), Apple Silicon (MPS), and AMD (DirectML/ROCm) with automatic CPU fallback, while caching embeddings locally to avoid re-indexing on updates. Features a tkinter GUI with drag-and-drop image search, negative term filtering, and background video indexing via FFmpeg.
Stars
65
Forks
7
Language
Python
License
—
Category
Last pushed
Mar 11, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/Ubaida-M-Yusuf/Makimus-AI"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
unum-cloud/UForm
Pocket-Sized Multimodal AI for content understanding and generation across multilingual texts,...
rom1504/clip-retrieval
Easily compute clip embeddings and build a clip retrieval system with them
mazzzystar/Queryable
Run OpenAI's CLIP and Apple's MobileCLIP model on iOS to search photos.
s-emanuilov/litepali
LitePali is a minimal, efficient implementation of ColPali for image retrieval and indexing,...
slavabarkov/tidy
Offline semantic Text-to-Image and Image-to-Image search on Android powered by quantized...