open_clip and simple-clip

These are competitors offering different trade-offs: open_clip is a production-ready, fully-featured CLIP implementation for practitioners needing robust performance and model variety, while simple-clip is a lightweight educational reference implementation optimized for understanding the core CLIP algorithm rather than practical deployment.

open_clip
86
Verified
simple-clip
54
Established
Maintenance 16/25
Adoption 25/25
Maturity 25/25
Community 20/25
Maintenance 0/25
Adoption 12/25
Maturity 25/25
Community 17/25
Stars: 13,496
Forks: 1,253
Downloads: 2,903,706
Commits (30d): 1
Language: Python
License:
Stars: 42
Forks: 8
Downloads: 39
Commits (30d): 0
Language: Jupyter Notebook
License: MIT
No risk flags
Stale 6m

About open_clip

mlfoundations/open_clip

An open source implementation of CLIP.

Supports diverse Vision Transformer and ConvNet architectures trained on large-scale datasets (LAION-2B, DataComp-1B) with published scaling laws, achieving competitive zero-shot ImageNet accuracy up to 85.4%. Integrates with PyTorch, Hugging Face model hub, and timm for image encoders, enabling efficient embedding computation via the clip-retrieval library. Offers flexible model loading from local checkpoints or HuggingFace, with pre-trained weights optimized for both inference and fine-tuning workflows.

About simple-clip

filipbasara0/simple-clip

A minimal, but effective implementation of CLIP (Contrastive Language-Image Pretraining) in PyTorch

Scores updated daily from GitHub, PyPI, and npm data. How scores work