StarlightSearch/EmbedAnything
Highly Performant, Modular, Memory Safe and Production-ready Inference, Ingestion and Indexing built in Rust 🦀
Supports multimodal ingestion (PDFs, images, audio) with pluggable vector database adapters and multiple embedding backends including Candle, ONNX, and cloud models. Uses Rust's memory-safe concurrency and streaming architecture to separate document preprocessing from inference across threads via MPSC channels, eliminating sequential bottlenecks while preventing memory leaks. Offers dense, sparse, and late-interaction embedding strategies with built-in semantic chunking methods for RAG workflows.
1,174 stars. Actively maintained with 3 commits in the last 30 days.
Stars
1,174
Forks
111
Language
Rust
License
Apache-2.0
Category
Last pushed
Mar 11, 2026
Commits (30d)
3
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/vector-db/StarlightSearch/EmbedAnything"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
databendlabs/databend
Data Agent Ready Warehouse : One for Analytics, Search, AI, Python Sandbox. — rebuilt from...
oceanbase/oceanbase
The Fastest Distributed Database for Transactional, Analytical, and AI Workloads.
matrixorigin/matrixone
MySQL-compatible HTAP database with Git for Data, vector search, and fulltext search....
ArcadeData/arcadedb
ArcadeDB Multi-Model Database, one DBMS that supports SQL, Cypher, Gremlin, HTTP/JSON, MongoDB...
lightonai/fast-plaid
High-Performance Engine for Multi-Vector Search