patrickjohncyh/fashion-clip
FashionCLIP is a CLIP-like model fine-tuned for the fashion domain.
Built on contrastive vision-language learning, FashionCLIP fine-tunes LAION CLIP checkpoints on 700K+ fashion image-text pairs from Farfetch, enabling zero-shot performance on domain-specific tasks like retrieval, classification, and attribute parsing. The model integrates with Hugging Face's transformers library and provides a dedicated Python API for efficient batch encoding of images and text into aligned embedding spaces, achieving significantly higher accuracy (F1 scores 0.62–0.83) across fashion benchmarks compared to general-purpose CLIP variants.
497 stars. No commits in the last 6 months.
Stars
497
Forks
52
Language
Python
License
MIT
Category
Last pushed
Jan 30, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/patrickjohncyh/fashion-clip"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ClipsAI/clipsai
Clips AI is an open-source Python library that automatically converts long videos into clips.
ai-forever/ru-clip
CLIP implementation for Russian language
Lednik7/CLIP-ONNX
It is a simple library to speed up CLIP inference up to 3x (K80 GPU)
suinleelab/CellCLIP
[NeurIPS 2025] CellCLIP – Learning Perturbation Effects in Cell Painting via Text-Guided...
cene555/ruCLIP-SB
RuCLIP-SB (Russian Contrastive Language–Image Pretraining SWIN-BERT) is a multimodal model for...