FlagEmbedding and mlx-embeddings

These are complements rather than competitors: FlagEmbedding provides a comprehensive retrieval and RAG framework suitable for cross-platform deployment, while MLX-Embeddings specializes in optimized inference for Mac-specific hardware (Apple Silicon via MLX), allowing users to run embedding models locally on macOS devices that would benefit from the retrieval capabilities FlagEmbedding provides.

FlagEmbedding
79
Verified
mlx-embeddings
69
Established
Maintenance 20/25
Adoption 15/25
Maturity 25/25
Community 19/25
Maintenance 10/25
Adoption 24/25
Maturity 18/25
Community 17/25
Stars: 11,395
Forks: 842
Downloads:
Commits (30d): 17
Language: Python
License: MIT
Stars: 290
Forks: 33
Downloads: 20,666
Commits (30d): 0
Language: Python
License:
No risk flags
No risk flags

About FlagEmbedding

FlagOpen/FlagEmbedding

Retrieval and Retrieval-augmented LLMs

Provides dense, sparse, and multi-vector embedding models (including BGE-M3 supporting 100+ languages and 8K context) alongside rerankers and multimodal variants for comprehensive semantic search and RAG pipelines. Built on transformer architectures with support for in-context learning, token compression, and unified retrieval methods—integrates seamlessly with vector databases and LLM frameworks via HuggingFace.

About mlx-embeddings

Blaizzy/mlx-embeddings

MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.

Supports multiple embedding model architectures (BERT, RoBERTa, ModernBERT, Qwen3-VL, Llama variants) and performs both unimodal text and multimodal text-image embedding generation via MLX framework. Provides batch processing capabilities with semantic similarity computation and task-specific functions like masked language modeling and sequence classification. Integrates with HuggingFace-compatible tokenizers and uses mean pooling or config-specified strategies for dense vector generation suitable for retrieval and reranking workflows.

Scores updated daily from GitHub, PyPI, and npm data. How scores work