patrickjohncyh/fashion-clip

FashionCLIP is a CLIP-like model fine-tuned for the fashion domain.

43
/ 100
Emerging

Built on contrastive vision-language learning, FashionCLIP fine-tunes LAION CLIP checkpoints on 700K+ fashion image-text pairs from Farfetch, enabling zero-shot performance on domain-specific tasks like retrieval, classification, and attribute parsing. The model integrates with Hugging Face's transformers library and provides a dedicated Python API for efficient batch encoding of images and text into aligned embedding spaces, achieving significantly higher accuracy (F1 scores 0.62–0.83) across fashion benchmarks compared to general-purpose CLIP variants.

497 stars. No commits in the last 6 months.

Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 17 / 25

How are scores calculated?

Stars

497

Forks

52

Language

Python

License

MIT

Last pushed

Jan 30, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/patrickjohncyh/fashion-clip"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.