FlagEmbedding and fastembed

These are complements—FlagEmbedding provides advanced embedding models and retrieval techniques, while FastEmbed provides the lightweight inference engine to efficiently run embedding models (including FlagEmbedding models) in production environments.

FlagEmbedding
79
Verified
fastembed
74
Verified
Maintenance 20/25
Adoption 15/25
Maturity 25/25
Community 19/25
Maintenance 16/25
Adoption 15/25
Maturity 25/25
Community 18/25
Stars: 11,395
Forks: 842
Downloads:
Commits (30d): 17
Language: Python
License: MIT
Stars: 2,771
Forks: 184
Downloads:
Commits (30d): 5
Language: Python
License: Apache-2.0
No risk flags
No risk flags

About FlagEmbedding

FlagOpen/FlagEmbedding

Retrieval and Retrieval-augmented LLMs

Provides dense, sparse, and multi-vector embedding models (including BGE-M3 supporting 100+ languages and 8K context) alongside rerankers and multimodal variants for comprehensive semantic search and RAG pipelines. Built on transformer architectures with support for in-context learning, token compression, and unified retrieval methods—integrates seamlessly with vector databases and LLM frameworks via HuggingFace.

About fastembed

qdrant/fastembed

Fast, Accurate, Lightweight Python library to make State of the Art Embedding

Leverages ONNX Runtime instead of PyTorch to minimize dependencies and enable deployment in serverless environments like AWS Lambda. Supports dense embeddings, sparse embeddings (SPLADE++), late-interaction models (ColBERT), image embeddings, and cross-encoder reranking—with extensibility for custom models. Integrates directly with Qdrant vector database for end-to-end semantic search workflows.

Scores updated daily from GitHub, PyPI, and npm data. How scores work